[ 465.544019] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.544467] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.544467] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.544795] env[62066]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 465.644067] env[62066]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62066) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 465.654529] env[62066]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62066) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 466.260388] env[62066]: INFO nova.virt.driver [None req-b2af7d5d-8609-4013-95ba-5696de25913f None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 466.334670] env[62066]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.334911] env[62066]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.334911] env[62066]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62066) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 469.462235] env[62066]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-5edec6e6-ff81-4f7d-beee-9bf0aa938b47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.478091] env[62066]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62066) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 469.478272] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-de09fdb0-7738-4434-93ee-56e203c085b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.514502] env[62066]: INFO oslo_vmware.api [-] Successfully established new session; session ID is ca666. [ 469.514701] env[62066]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.180s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.515195] env[62066]: INFO nova.virt.vmwareapi.driver [None req-b2af7d5d-8609-4013-95ba-5696de25913f None None] VMware vCenter version: 7.0.3 [ 469.518669] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f081f8-3f6a-466c-9402-597d9a389ac1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.539297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20cd4b0-125d-468d-a111-ef9bfe622237 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.544996] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968cb9f2-2641-4e68-a38f-9c13f6ed1328 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.551483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a51f5d1-0879-4659-a29d-970a0c588b1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.564618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85072459-db08-4a67-869b-8d35c9c1bf29 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.570378] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d65bd1d-6797-46cd-a73f-06607791cd08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.600121] env[62066]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-bc84cce4-f52f-48ab-9409-a5bdb55c495b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.604840] env[62066]: DEBUG nova.virt.vmwareapi.driver [None req-b2af7d5d-8609-4013-95ba-5696de25913f None None] Extension org.openstack.compute already exists. {{(pid=62066) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 469.607514] env[62066]: INFO nova.compute.provider_config [None req-b2af7d5d-8609-4013-95ba-5696de25913f None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 470.111178] env[62066]: DEBUG nova.context [None req-b2af7d5d-8609-4013-95ba-5696de25913f None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),265cf9d3-88f6-4655-aa47-be12f115d030(cell1) {{(pid=62066) load_cells /opt/stack/nova/nova/context.py:464}} [ 470.113236] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.113461] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.114141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.114589] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Acquiring lock "265cf9d3-88f6-4655-aa47-be12f115d030" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.114769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Lock "265cf9d3-88f6-4655-aa47-be12f115d030" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.115905] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Lock "265cf9d3-88f6-4655-aa47-be12f115d030" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.136349] env[62066]: INFO dbcounter [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Registered counter for database nova_cell0 [ 470.144616] env[62066]: INFO dbcounter [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Registered counter for database nova_cell1 [ 470.148091] env[62066]: DEBUG oslo_db.sqlalchemy.engines [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62066) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.148468] env[62066]: DEBUG oslo_db.sqlalchemy.engines [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62066) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.153543] env[62066]: ERROR nova.db.main.api [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.153543] env[62066]: result = function(*args, **kwargs) [ 470.153543] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 470.153543] env[62066]: return func(*args, **kwargs) [ 470.153543] env[62066]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.153543] env[62066]: result = fn(*args, **kwargs) [ 470.153543] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.153543] env[62066]: return f(*args, **kwargs) [ 470.153543] env[62066]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.153543] env[62066]: return db.service_get_minimum_version(context, binaries) [ 470.153543] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.153543] env[62066]: _check_db_access() [ 470.153543] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.153543] env[62066]: stacktrace = ''.join(traceback.format_stack()) [ 470.153543] env[62066]: [ 470.154255] env[62066]: ERROR nova.db.main.api [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.154255] env[62066]: result = function(*args, **kwargs) [ 470.154255] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 470.154255] env[62066]: return func(*args, **kwargs) [ 470.154255] env[62066]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.154255] env[62066]: result = fn(*args, **kwargs) [ 470.154255] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.154255] env[62066]: return f(*args, **kwargs) [ 470.154255] env[62066]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.154255] env[62066]: return db.service_get_minimum_version(context, binaries) [ 470.154255] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.154255] env[62066]: _check_db_access() [ 470.154255] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.154255] env[62066]: stacktrace = ''.join(traceback.format_stack()) [ 470.154255] env[62066]: [ 470.154837] env[62066]: WARNING nova.objects.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Failed to get minimum service version for cell 265cf9d3-88f6-4655-aa47-be12f115d030 [ 470.154837] env[62066]: WARNING nova.objects.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 470.155212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Acquiring lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 470.155389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Acquired lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 470.155658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Releasing lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 470.155994] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Full set of CONF: {{(pid=62066) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 470.156156] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ******************************************************************************** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 470.156288] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Configuration options gathered from: {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 470.156432] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 470.156630] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 470.156760] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ================================================================================ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 470.156969] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] allow_resize_to_same_host = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157153] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] arq_binding_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157288] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] backdoor_port = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157414] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] backdoor_socket = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157612] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] block_device_allocate_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157784] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] block_device_allocate_retries_interval = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.157950] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cert = self.pem {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158128] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158299] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute_monitors = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158480] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] config_dir = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158663] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] config_drive_format = iso9660 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158798] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.158964] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] config_source = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159151] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] console_host = devstack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159319] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] control_exchange = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159476] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cpu_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159636] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] daemon = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159803] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] debug = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.159957] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_access_ip_network_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160135] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_availability_zone = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160293] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_ephemeral_format = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160449] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_green_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160681] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160842] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] default_schedule_zone = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.160995] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] disk_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161165] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] enable_new_services = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161340] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] enabled_apis = ['osapi_compute'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161502] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] enabled_ssl_apis = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161663] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] flat_injected = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161845] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] force_config_drive = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.161971] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] force_raw_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.162152] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] graceful_shutdown_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.162310] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] heal_instance_info_cache_interval = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.162523] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] host = cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.162826] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163014] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] initial_disk_allocation_ratio = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163186] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] initial_ram_allocation_ratio = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163402] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163568] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_build_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163732] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_delete_interval = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.163900] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_format = [instance: %(uuid)s] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164078] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_name_template = instance-%08x {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164245] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_usage_audit = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164417] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_usage_audit_period = month {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164583] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164750] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] instances_path = /opt/stack/data/nova/instances {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.164915] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] internal_service_availability_zone = internal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165083] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] key = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165246] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] live_migration_retry_count = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165413] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_color = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165608] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_config_append = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165783] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.165943] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_dir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166117] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166252] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_options = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166415] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_rotate_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166584] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_rotate_interval_type = days {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166753] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] log_rotation_type = none {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.166893] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167033] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167207] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167371] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167523] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167701] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] long_rpc_timeout = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.167867] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_concurrent_builds = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168034] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_concurrent_live_migrations = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168201] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_concurrent_snapshots = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168360] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_local_block_devices = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168539] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_logfile_count = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168709] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] max_logfile_size_mb = 200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.168871] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] maximum_instance_delete_attempts = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169049] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metadata_listen = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169222] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metadata_listen_port = 8775 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169392] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metadata_workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169554] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] migrate_max_retries = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169722] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] mkisofs_cmd = genisoimage {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.169929] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] my_block_storage_ip = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170074] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] my_ip = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170241] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] network_allocate_retries = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170418] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170584] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] osapi_compute_listen = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170747] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] osapi_compute_listen_port = 8774 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.170920] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] osapi_compute_unique_server_name_scope = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171094] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] osapi_compute_workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171261] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] password_length = 12 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171420] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] periodic_enable = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171579] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] periodic_fuzzy_delay = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171750] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] pointer_model = usbtablet {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.171918] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] preallocate_images = none {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172087] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] publish_errors = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172220] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] pybasedir = /opt/stack/nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172373] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ram_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172530] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rate_limit_burst = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172776] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rate_limit_except_level = CRITICAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.172958] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rate_limit_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173136] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reboot_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173297] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reclaim_instance_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173454] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] record = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173621] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reimage_timeout_per_gb = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173787] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] report_interval = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.173948] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rescue_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174120] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reserved_host_cpus = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174278] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reserved_host_disk_mb = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174435] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reserved_host_memory_mb = 512 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174591] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] reserved_huge_pages = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174787] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] resize_confirm_window = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.174954] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] resize_fs_using_block_device = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175126] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] resume_guests_state_on_host_boot = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175292] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175458] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] rpc_response_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175636] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] run_external_periodic_tasks = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175804] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] running_deleted_instance_action = reap {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.175963] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] running_deleted_instance_poll_interval = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176134] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] running_deleted_instance_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176291] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler_instance_sync_interval = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176456] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_down_time = 720 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176622] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] servicegroup_driver = db {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176774] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] shell_completion = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.176929] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] shelved_offload_time = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.177098] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] shelved_poll_interval = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.177266] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] shutdown_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.177422] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] source_is_ipv6 = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.177615] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ssl_only = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.177867] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178047] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] sync_power_state_interval = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178214] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] sync_power_state_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178381] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] syslog_log_facility = LOG_USER {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178560] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] tempdir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178729] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] timeout_nbd = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.178895] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] transport_url = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179067] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] update_resources_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179230] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_cow_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179387] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_eventlog = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179542] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_journal = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179701] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_json = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.179857] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_rootwrap_daemon = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180023] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_stderr = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180185] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] use_syslog = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180341] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vcpu_pin_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180505] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plugging_is_fatal = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180670] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plugging_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180831] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] virt_mkfs = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.180988] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] volume_usage_poll_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.181160] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] watch_log_file = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.181327] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] web = /usr/share/spice-html5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.181510] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.181676] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.181840] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.182014] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_concurrency.disable_process_locking = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.182555] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.182755] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183038] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183229] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183405] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183575] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183761] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.auth_strategy = keystone {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.183931] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.compute_link_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184122] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184301] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.dhcp_domain = novalocal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184472] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.enable_instance_password = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184643] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.glance_link_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184811] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.184979] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.185158] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.instance_list_per_project_cells = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.185321] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.list_records_by_skipping_down_cells = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.185505] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.local_metadata_per_cell = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.185696] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.max_limit = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.185867] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.metadata_cache_expiration = 15 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186053] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.neutron_default_tenant_id = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186231] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.response_validation = warn {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186405] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.use_neutron_default_nets = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186574] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186770] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.186956] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.187147] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.187322] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_dynamic_targets = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.187513] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_jsonfile_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.187705] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.187900] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.backend = dogpile.cache.memcached {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188080] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.backend_argument = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188258] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.config_prefix = cache.oslo {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188429] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.dead_timeout = 60.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188593] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.debug_cache_backend = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188755] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.enable_retry_client = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.188915] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.enable_socket_keepalive = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189096] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.enabled = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189264] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.enforce_fips_mode = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189429] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.expiration_time = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189594] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.hashclient_retry_attempts = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189764] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.hashclient_retry_delay = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.189928] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_dead_retry = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190097] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190263] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190423] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190586] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_pool_maxsize = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190747] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.190908] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_sasl_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191094] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191264] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_socket_timeout = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191423] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.memcache_username = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191588] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.proxies = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191753] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_db = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.191912] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.192092] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_sentinel_service_name = mymaster {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.192270] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.192441] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_server = localhost:6379 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.192607] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_socket_timeout = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.192765] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.redis_username = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193014] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.retry_attempts = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193206] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.retry_delay = 0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193377] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.socket_keepalive_count = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193542] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.socket_keepalive_idle = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193707] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.socket_keepalive_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.193866] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.tls_allowed_ciphers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194033] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.tls_cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194198] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.tls_certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194360] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.tls_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194520] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cache.tls_keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194692] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.194864] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195035] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195214] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.catalog_info = volumev3::publicURL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195377] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195542] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195709] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.cross_az_attach = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.195872] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196041] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.endpoint_template = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196212] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.http_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196375] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196534] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196707] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.os_region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.196871] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197039] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cinder.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197215] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197377] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.cpu_dedicated_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197540] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.cpu_shared_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197707] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.image_type_exclude_list = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.197869] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198041] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.max_concurrent_disk_ops = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198209] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.max_disk_devices_to_attach = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198371] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198571] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198749] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.resource_provider_association_refresh = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.198914] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199090] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.shutdown_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199276] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199455] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] conductor.workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199636] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] console.allowed_origins = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199797] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] console.ssl_ciphers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.199967] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] console.ssl_minimum_version = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200151] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] consoleauth.enforce_session_timeout = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200321] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] consoleauth.token_ttl = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200490] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200649] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200814] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.200973] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201143] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201303] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201465] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201626] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201786] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.201943] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202119] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202281] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202441] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202611] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.service_type = accelerator {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202773] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.202930] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.203188] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.203361] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.203547] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.203712] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] cyborg.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.203894] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.backend = sqlalchemy {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204079] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204250] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.connection_debug = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204421] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.connection_parameters = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204587] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.connection_recycle_time = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204752] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.connection_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.204915] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.db_inc_retry_interval = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205090] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.db_max_retries = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205260] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.db_max_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205425] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.db_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205586] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.max_overflow = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205748] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.max_pool_size = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.205908] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.max_retries = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206087] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206249] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.mysql_wsrep_sync_wait = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206406] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.pool_timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206567] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206724] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.slave_connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.206884] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.sqlite_synchronous = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207053] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] database.use_db_reconnect = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207234] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.backend = sqlalchemy {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207402] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207596] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.connection_debug = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207802] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.connection_parameters = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.207971] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.connection_recycle_time = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.208149] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.connection_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.208313] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.db_inc_retry_interval = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.208508] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.db_max_retries = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.208709] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.db_max_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.208880] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.db_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209059] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.max_overflow = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209228] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.max_pool_size = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209388] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.max_retries = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209559] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209720] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.209877] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.pool_timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210047] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210209] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.slave_connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210369] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] api_database.sqlite_synchronous = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210564] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] devices.enabled_mdev_types = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210762] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.210980] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ephemeral_storage_encryption.default_format = luks {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.211201] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ephemeral_storage_encryption.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.211389] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.211565] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.api_servers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.211732] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.211896] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212073] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212239] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212400] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212589] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212775] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.default_trusted_certificate_ids = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.212993] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.enable_certificate_validation = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.213286] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.enable_rbd_download = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.213471] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.213648] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.213815] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.213979] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.214157] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.214324] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.num_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.214495] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.rbd_ceph_conf = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.214661] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.rbd_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.214831] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.rbd_pool = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.215037] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.rbd_user = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.215218] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.215446] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.215642] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.215859] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.service_type = image {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216042] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216215] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216376] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216537] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216722] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.216956] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.verify_glance_signatures = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.217144] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] glance.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.217316] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] guestfs.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.217519] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] mks.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.217875] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.218080] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.manager_interval = 2400 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.218257] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.precache_concurrency = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.218506] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.remove_unused_base_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.218701] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.218876] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219066] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] image_cache.subdirectory_name = _base {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219250] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.api_max_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219417] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.api_retry_interval = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219580] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219746] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.219904] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.220074] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.220302] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.220536] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.conductor_group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.220696] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.220871] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221045] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221217] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221379] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221538] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221698] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.221865] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.peer_list = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222031] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222194] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222357] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.serial_console_state_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222515] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222688] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.service_type = baremetal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.222841] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.shard = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223009] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223173] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223425] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223591] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223776] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.223937] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ironic.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224132] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224311] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] key_manager.fixed_key = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224491] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224654] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.barbican_api_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224811] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.barbican_endpoint = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.224983] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.barbican_endpoint_type = public {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225155] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.barbican_region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225313] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225471] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225632] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225791] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.225945] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226122] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.number_of_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226287] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.retry_delay = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226448] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.send_service_user_token = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226612] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226767] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.226925] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.verify_ssl = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227093] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican.verify_ssl_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227263] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227424] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227615] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227779] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.227943] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228117] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228277] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228467] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228651] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] barbican_service_user.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228820] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.approle_role_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.228980] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.approle_secret_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229162] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.kv_mountpoint = secret {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229320] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.kv_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229481] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.kv_version = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229638] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.namespace = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229795] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.root_token_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.229953] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.ssl_ca_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230134] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.timeout = 60.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230299] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.use_ssl = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230467] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230642] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230804] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.230961] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231131] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231293] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231473] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231649] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231808] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.231970] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232149] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232334] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232496] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232656] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232816] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.232974] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.233157] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.service_type = identity {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.233387] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.233574] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.233739] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.233913] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.234136] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.234308] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] keystone.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.234533] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.connection_uri = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.234712] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_mode = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.234881] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_model_extra_flags = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235068] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_models = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235247] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_power_governor_high = performance {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235419] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_power_governor_low = powersave {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235583] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_power_management = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235756] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.235921] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.device_detach_attempts = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236095] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.device_detach_timeout = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236267] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.disk_cachemodes = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236426] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.disk_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236593] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.enabled_perf_events = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236757] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.file_backed_memory = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.236921] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.gid_maps = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237090] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.hw_disk_discard = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237251] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.hw_machine_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237421] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_rbd_ceph_conf = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237630] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237807] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.237979] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_rbd_glance_store_name = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.238166] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_rbd_pool = rbd {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.238341] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_type = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.238519] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.images_volume_group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.238695] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.inject_key = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.238859] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.inject_partition = -2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239031] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.inject_password = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239202] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.iscsi_iface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239366] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.iser_use_multipath = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239532] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_bandwidth = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239696] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.239860] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_downtime = 500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240032] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240202] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240363] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_inbound_addr = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240537] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240714] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_permit_post_copy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.240888] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_scheme = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241107] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_timeout_action = abort {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241281] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_tunnelled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241444] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241610] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.live_migration_with_native_tls = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241770] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.max_queues = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.241934] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.242193] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.242362] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.nfs_mount_options = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.242651] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.242826] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.242991] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_iser_scan_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.243168] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_memory_encrypted_guests = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.243333] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.243583] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_pcie_ports = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.243769] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.num_volume_scan_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.243939] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.pmem_namespaces = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.244115] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.quobyte_client_cfg = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.244433] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.244615] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rbd_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.244786] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.244953] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245129] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rbd_secret_uuid = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245294] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rbd_user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245460] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245636] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.remote_filesystem_transport = ssh {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245798] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rescue_image_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.245955] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rescue_kernel_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.246128] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rescue_ramdisk_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.246299] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.246462] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.rx_queue_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.246630] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.smbfs_mount_options = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.246901] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.247088] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.snapshot_compression = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.247255] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.snapshot_image_format = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.247493] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.247678] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.sparse_logical_volumes = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.247845] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.swtpm_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248027] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.swtpm_group = tss {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248205] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.swtpm_user = tss {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248377] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.sysinfo_serial = unique {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248565] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.tb_cache_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248735] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.tx_queue_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.248903] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.uid_maps = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249115] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.use_virtio_for_bridges = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249292] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.virt_type = kvm {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249462] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.volume_clear = zero {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249629] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.volume_clear_size = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249795] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.volume_use_multipath = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.249953] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_cache_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.250210] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.250389] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_mount_group = qemu {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.250557] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_mount_opts = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.250728] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251011] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251199] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.vzstorage_mount_user = stack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251368] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251546] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251723] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.251883] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252050] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252217] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252373] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252529] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252699] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.default_floating_pool = public {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.252855] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253024] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.extension_sync_interval = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253189] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.http_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253347] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253503] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253755] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.253934] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254107] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254279] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.ovs_bridge = br-int {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254444] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.physnets = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254611] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254770] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.254937] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.service_metadata_proxy = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255107] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255280] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.service_type = network {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255446] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255606] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255762] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.255918] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256108] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256271] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] neutron.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256442] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] notifications.bdms_in_notifications = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256620] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] notifications.default_level = INFO {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256793] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] notifications.notification_format = unversioned {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.256956] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] notifications.notify_on_state_change = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.257145] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.257323] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] pci.alias = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.257514] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] pci.device_spec = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.257694] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] pci.report_in_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.257868] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258049] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258220] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258379] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258561] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258731] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.258888] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259053] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259215] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.default_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259370] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.default_domain_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259525] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259683] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.domain_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259836] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.259994] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260161] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260316] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260470] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260638] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260797] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.project_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.260961] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.project_domain_name = Default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261135] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.project_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261308] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.project_name = service {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261474] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261638] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261795] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.261957] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.service_type = placement {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262132] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262291] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262447] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262604] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.system_scope = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262759] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.262913] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.trust_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.263079] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.user_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.263250] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.user_domain_name = Default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.263407] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.user_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.263578] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.username = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.263849] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264029] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] placement.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264213] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.cores = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264378] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.count_usage_from_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264549] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264727] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.injected_file_content_bytes = 10240 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.264892] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.injected_file_path_length = 255 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265068] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.injected_files = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265235] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.instances = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265398] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.key_pairs = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265560] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.metadata_items = 128 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265724] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.ram = 51200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.265887] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.recheck_quota = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266065] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.server_group_members = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266237] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] quota.server_groups = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266414] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266579] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266747] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.image_metadata_prefilter = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.266911] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267086] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.max_attempts = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267253] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.max_placement_results = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267416] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267605] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.query_placement_for_image_type_support = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267773] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.267947] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] scheduler.workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.268134] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.268309] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.268511] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.268694] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.268860] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269031] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269204] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269396] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269566] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.host_subset_size = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269732] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.269891] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270063] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270232] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.isolated_hosts = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270396] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.isolated_images = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270556] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270717] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.270886] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271058] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.pci_in_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271225] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271384] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271544] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271703] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.271862] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272030] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272194] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.track_instance_changes = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272368] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272537] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metrics.required = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272700] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metrics.weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.272860] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.273027] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] metrics.weight_setting = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.273341] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.273516] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.273691] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.port_range = 10000:20000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.273950] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274143] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274314] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] serial_console.serialproxy_port = 6083 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274482] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274656] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274817] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.274974] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275154] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275315] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275472] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275642] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.send_service_user_token = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275802] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.275973] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] service_user.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.276159] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.agent_enabled = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.276323] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.276628] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.276822] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.276991] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.html5proxy_port = 6082 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277165] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.image_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277322] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.jpeg_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277484] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.playback_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277648] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.require_secure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277813] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.server_listen = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.277979] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278150] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.streaming_mode = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278307] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] spice.zlib_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278473] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] upgrade_levels.baseapi = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278641] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] upgrade_levels.compute = auto {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278798] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] upgrade_levels.conductor = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.278953] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] upgrade_levels.scheduler = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279134] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279297] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279455] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279615] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279777] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.279933] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280103] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280265] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280422] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vendordata_dynamic_auth.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280595] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.api_retry_count = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280759] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.280924] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.cache_prefix = devstack-image-cache {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281102] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.cluster_name = testcl1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281270] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.connection_pool_size = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281431] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.console_delay_seconds = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281599] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.datastore_regex = ^datastore.* {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281805] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.281975] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.host_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282157] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.host_port = 443 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282326] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.host_username = administrator@vsphere.local {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282494] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.insecure = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282658] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.integration_bridge = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282820] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.maximum_objects = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.282977] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.pbm_default_policy = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.283154] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.pbm_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.283311] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.pbm_wsdl_location = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.283476] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.283635] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.serial_port_proxy_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.283790] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.serial_port_service_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284053] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.task_poll_interval = 0.5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284236] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.use_linked_clone = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284407] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.vnc_keymap = en-us {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284574] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.vnc_port = 5900 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284737] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vmware.vnc_port_total = 10000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.284919] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.auth_schemes = ['none'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.285108] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.285404] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.285588] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.285760] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.novncproxy_port = 6080 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.285936] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.server_listen = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286122] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286285] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.vencrypt_ca_certs = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286444] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.vencrypt_client_cert = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286601] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vnc.vencrypt_client_key = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286778] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.286940] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_deep_image_inspection = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287113] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287275] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287437] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287602] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.disable_rootwrap = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287764] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.enable_numa_live_migration = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.287922] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.288569] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.288759] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.288931] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.libvirt_disable_apic = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289110] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289279] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289446] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289617] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289778] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.289943] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290118] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290283] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290445] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290614] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290800] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.290973] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.client_socket_timeout = 900 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291160] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.default_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291331] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.keep_alive = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291501] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.max_header_line = 16384 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291666] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.secure_proxy_ssl_header = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291830] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.ssl_ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.291990] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.ssl_cert_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.292163] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.ssl_key_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.292326] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.tcp_keepidle = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.292503] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.292673] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] zvm.ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.292834] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] zvm.cloud_connector_url = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.293139] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.293313] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] zvm.reachable_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.293495] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.enforce_new_defaults = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.293898] env[62066]: WARNING oslo_config.cfg [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 470.294094] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.enforce_scope = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.294278] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.policy_default_rule = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.294542] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.294747] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.policy_file = policy.yaml {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.294931] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295114] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295278] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295440] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295602] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295772] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.295949] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296142] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.connection_string = messaging:// {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296315] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296489] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.es_doc_type = notification {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296658] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.es_scroll_size = 10000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296829] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.es_scroll_time = 2m {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.296993] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.filter_error_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.297179] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.hmac_keys = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.297360] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.sentinel_service_name = mymaster {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.297544] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.socket_timeout = 0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.297697] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.trace_requests = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.297867] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler.trace_sqlalchemy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298065] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler_jaeger.process_tags = {} {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298233] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler_jaeger.service_name_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298399] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] profiler_otlp.service_name_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298566] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] remote_debug.host = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298727] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] remote_debug.port = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.298907] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299082] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299250] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299412] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299574] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299752] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.299914] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300086] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300250] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300420] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300580] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300756] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.300924] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301111] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301319] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301482] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301650] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301823] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.301984] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302161] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302328] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302492] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302654] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302821] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.302983] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303161] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303324] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303486] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303657] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303824] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.303994] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.304178] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.304343] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.304587] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.304789] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.ssl_version = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.304961] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.305166] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.305338] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_notifications.retry = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.305522] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.305701] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_messaging_notifications.transport_url = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.305884] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306064] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306229] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306389] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306553] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306713] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.306872] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307040] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.endpoint_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307210] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307371] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307532] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307815] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307871] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.307995] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.308170] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.308361] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.308523] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.service_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.308688] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.308845] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309008] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309175] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309335] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.valid_interfaces = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309491] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_limit.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309658] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_reports.file_event_handler = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309822] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.309980] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] oslo_reports.log_dir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310167] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310329] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310487] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310655] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310817] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.310975] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311161] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311323] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311483] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311650] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311812] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.311969] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] vif_plug_ovs_privileged.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.312154] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.flat_interface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.312336] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.312511] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.312685] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.312855] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313032] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313206] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313370] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313548] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313721] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.isolate_vif = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.313891] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.314070] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.314242] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.314412] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.ovsdb_interface = native {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.314577] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] os_vif_ovs.per_port_bridge = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.314843] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.capabilities = [21] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315018] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315185] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315355] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315518] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315680] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] privsep_osbrick.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.315853] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316022] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316187] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316387] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316559] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316721] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] nova_sys_admin.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.316854] env[62066]: DEBUG oslo_service.service [None req-d5ffd63a-b980-413b-8d6b-23e4cc76ad3b None None] ******************************************************************************** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 470.317397] env[62066]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 470.820505] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Getting list of instances from cluster (obj){ [ 470.820505] env[62066]: value = "domain-c8" [ 470.820505] env[62066]: _type = "ClusterComputeResource" [ 470.820505] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.821699] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d882cd1-887e-4c03-b963-1858265d638f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.830818] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Got total of 0 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 470.831433] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-d0b1961e-0817-4255-a947-53e546015907 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 470.831913] env[62066]: INFO nova.virt.node [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Generated node identity cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 [ 470.832169] env[62066]: INFO nova.virt.node [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Wrote node identity cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 to /opt/stack/data/n-cpu-1/compute_id [ 471.336054] env[62066]: WARNING nova.compute.manager [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Compute nodes ['cd4c0e36-9c88-4f73-a93c-1ff383ed97c4'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 472.342090] env[62066]: INFO nova.compute.manager [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 473.350026] env[62066]: WARNING nova.compute.manager [None req-d0b1961e-0817-4255-a947-53e546015907 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 473.350026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.350026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.350026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 473.350026] env[62066]: DEBUG nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 473.350026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95cfaab-3cdc-4aec-9d06-d85c60b80ee1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.359209] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91349a53-ee6e-4532-a275-858a0fd187e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.373726] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f703767-1c12-4993-b44b-2ef9f4ed63ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.380331] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86d9430-acac-4d99-b822-e6766df9ab68 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.410519] env[62066]: DEBUG nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181538MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 473.410889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.411247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.915991] env[62066]: WARNING nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] No compute node record for cpu-1:cd4c0e36-9c88-4f73-a93c-1ff383ed97c4: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 could not be found. [ 474.419541] env[62066]: INFO nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 [ 475.930324] env[62066]: DEBUG nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 475.930324] env[62066]: DEBUG nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 476.080585] env[62066]: INFO nova.scheduler.client.report [None req-d0b1961e-0817-4255-a947-53e546015907 None None] [req-a61637fc-0ea5-4790-b708-c437fda866c0] Created resource provider record via placement API for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 476.096086] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40834aa7-d1e6-41c7-902f-700eca69d4e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.103378] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2afa7f5-1162-4ba5-8121-07d919153658 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.136049] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03343194-66ec-4438-b447-6ecd03dfb942 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.142301] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8097c233-38ef-44f3-8555-74091394b861 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.158275] env[62066]: DEBUG nova.compute.provider_tree [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.701022] env[62066]: DEBUG nova.scheduler.client.report [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 476.701022] env[62066]: DEBUG nova.compute.provider_tree [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 0 to 1 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.701022] env[62066]: DEBUG nova.compute.provider_tree [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.746893] env[62066]: DEBUG nova.compute.provider_tree [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 1 to 2 during operation: update_traits {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.254287] env[62066]: DEBUG nova.compute.resource_tracker [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 477.254287] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.841s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.254287] env[62066]: DEBUG nova.service [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Creating RPC server for service compute {{(pid=62066) start /opt/stack/nova/nova/service.py:186}} [ 477.267266] env[62066]: DEBUG nova.service [None req-d0b1961e-0817-4255-a947-53e546015907 None None] Join ServiceGroup membership for this service compute {{(pid=62066) start /opt/stack/nova/nova/service.py:203}} [ 477.268271] env[62066]: DEBUG nova.servicegroup.drivers.db [None req-d0b1961e-0817-4255-a947-53e546015907 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62066) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 513.890087] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "6eea6917-0ed4-447a-96f0-094c1832160a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.890087] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "6eea6917-0ed4-447a-96f0-094c1832160a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.900243] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.900345] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.394789] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.407540] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.961420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.961809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.964065] env[62066]: INFO nova.compute.claims [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.969523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.911200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquiring lock "5b78d0ac-2874-417a-ad2d-d4197d6fbe95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.913031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "5b78d0ac-2874-417a-ad2d-d4197d6fbe95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.071111] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0102d9-3614-4955-b876-1f72b4a1a4f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.087429] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2543c6-f9e8-4d75-9b3a-c799108f46aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.132965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaa69e1-1daf-4f23-922d-3f99a76a0c24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.141378] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee0da1d-c9e7-4cde-af28-f59a007c816d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.157667] env[62066]: DEBUG nova.compute.provider_tree [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.418701] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.655396] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "cdccd327-293e-4401-905c-4c16150e493a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.655688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "cdccd327-293e-4401-905c-4c16150e493a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.664267] env[62066]: DEBUG nova.scheduler.client.report [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.959189] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.161248] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.175677] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.176772] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.179950] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.210s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.182027] env[62066]: INFO nova.compute.claims [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.621166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquiring lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.621458] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.691207] env[62066]: DEBUG nova.compute.utils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 517.694119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.694428] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.694709] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 518.124381] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.173398] env[62066]: DEBUG nova.policy [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '340f5d349bf848919d32407ed28d9949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a079fb2263e4444880bd8f56fa0e28eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 518.200693] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.333858] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606e1f2c-7909-490c-8d44-aa2babe050f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.341985] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5801b5-d12e-4d71-8dbd-ff537e50c988 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.380884] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af28e5e-b978-400b-ad94-9bed0fad05db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.388949] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef423b5-ef4f-4490-bcb5-311295725760 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.403741] env[62066]: DEBUG nova.compute.provider_tree [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.659769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.732822] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Successfully created port: 26bc6c6c-abcb-4f06-b02e-020492ead1c7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.906656] env[62066]: DEBUG nova.scheduler.client.report [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.213405] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.244306] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.244306] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.244306] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.244475] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.244475] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.244475] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.244666] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.244815] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.246092] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.246301] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.246502] env[62066]: DEBUG nova.virt.hardware [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.247928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501f978d-c79d-4f38-8d16-fd6a48a87309 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.258389] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183a1ffc-1519-47ff-9e6c-674e33499c9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.286020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3699d941-022e-4302-86a0-8743076ff839 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.415169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.415575] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.422151] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.460s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.422151] env[62066]: INFO nova.compute.claims [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.925140] env[62066]: DEBUG nova.compute.utils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 519.929180] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 519.929180] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 520.089717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.089717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.096193] env[62066]: DEBUG nova.policy [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48182fb845b476c9cdf15dc99f9c1da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae83fb1cd0cf41c2ae932fa5d06ea4ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.430371] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.561406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753f06dd-f0bc-4f49-b4ee-d8fc14625e4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.571549] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbda2bdc-1199-47ea-bd83-b445530ed06e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.602552] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.607043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f9932d-ffb0-4a94-ae70-3847f047c8a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.616177] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e40bff4-0ca4-4884-94db-ea195f984435 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.631524] env[62066]: DEBUG nova.compute.provider_tree [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.136281] env[62066]: DEBUG nova.scheduler.client.report [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.149383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.274073] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Successfully created port: 30270b72-5e44-481c-9660-885918d4c9e4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.442315] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 521.474039] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 521.474278] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 521.474496] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 521.474855] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 521.475049] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 521.475318] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 521.475666] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 521.475666] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 521.476016] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 521.476016] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 521.476272] env[62066]: DEBUG nova.virt.hardware [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 521.477086] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3d0633-b898-4f34-997f-ddf5cbdc246a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.486601] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f8312a-5d72-4bf6-bd3e-e4b08f515588 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.644966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.227s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.645538] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.654070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.954s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.654070] env[62066]: INFO nova.compute.claims [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.158765] env[62066]: DEBUG nova.compute.utils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 522.163163] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.163356] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 522.329448] env[62066]: DEBUG nova.policy [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '49d3f1671c3a4dcda83419799ad5f45b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ab2bed68b0a421ba994aae759068d88', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.664717] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.808031] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04032336-6fcf-4d0a-84f8-1b34c25b3c9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.816533] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9fb7f0-d73a-46c6-94c8-3fa5ef68d701 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.865700] env[62066]: ERROR nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 522.865700] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.865700] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.865700] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.865700] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.865700] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.865700] env[62066]: ERROR nova.compute.manager raise self.value [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.865700] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 522.865700] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.865700] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 522.866187] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.866187] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 522.866187] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 522.866187] env[62066]: ERROR nova.compute.manager [ 522.866187] env[62066]: Traceback (most recent call last): [ 522.866187] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 522.866187] env[62066]: listener.cb(fileno) [ 522.866187] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.866187] env[62066]: result = function(*args, **kwargs) [ 522.866187] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.866187] env[62066]: return func(*args, **kwargs) [ 522.866187] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.866187] env[62066]: raise e [ 522.866187] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.866187] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 522.866187] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.866187] env[62066]: created_port_ids = self._update_ports_for_instance( [ 522.866187] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.866187] env[62066]: with excutils.save_and_reraise_exception(): [ 522.866187] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.866187] env[62066]: self.force_reraise() [ 522.866187] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.866187] env[62066]: raise self.value [ 522.866187] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.866187] env[62066]: updated_port = self._update_port( [ 522.866187] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.866187] env[62066]: _ensure_no_port_binding_failure(port) [ 522.866187] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.866187] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 522.866969] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 522.866969] env[62066]: Removing descriptor: 16 [ 522.868120] env[62066]: ERROR nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Traceback (most recent call last): [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] yield resources [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.driver.spawn(context, instance, image_meta, [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] vm_ref = self.build_virtual_machine(instance, [ 522.868120] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] for vif in network_info: [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self._sync_wrapper(fn, *args, **kwargs) [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.wait() [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self[:] = self._gt.wait() [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self._exit_event.wait() [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.868580] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] result = hub.switch() [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self.greenlet.switch() [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] result = function(*args, **kwargs) [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return func(*args, **kwargs) [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise e [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] nwinfo = self.network_api.allocate_for_instance( [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] created_port_ids = self._update_ports_for_instance( [ 522.868922] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] with excutils.save_and_reraise_exception(): [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.force_reraise() [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise self.value [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] updated_port = self._update_port( [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] _ensure_no_port_binding_failure(port) [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise exception.PortBindingFailed(port_id=port['id']) [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 522.869257] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] [ 522.869681] env[62066]: INFO nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Terminating instance [ 522.870436] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d6e31f-b67b-4964-9640-66b013d6cc04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.873338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.873459] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.873619] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 522.881625] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bf589b-11e5-436e-9612-75a5bca4833a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.896061] env[62066]: DEBUG nova.compute.provider_tree [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.305969] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Successfully created port: 4866db3b-9a37-48d8-9bb8-3127ce1f124e {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.398826] env[62066]: DEBUG nova.scheduler.client.report [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.453229] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 523.682067] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.712385] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.712649] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.712807] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.712983] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.713143] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.713290] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.713491] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.713647] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.713814] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.713964] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.714545] env[62066]: DEBUG nova.virt.hardware [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.715633] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e091ff-dd8f-4b82-a64d-41dc1d191f40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.730676] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae6b5a1-6fc3-4563-83f3-e5b73421f3c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.904168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.904939] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.907909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.249s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.913722] env[62066]: INFO nova.compute.claims [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.930304] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.231141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "6e2496af-4d60-4315-9816-d0456458bc13" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.231682] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "6e2496af-4d60-4315-9816-d0456458bc13" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.420199] env[62066]: DEBUG nova.compute.utils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 524.424741] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 524.424929] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 524.436130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.436130] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.436130] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.436130] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-244a1b84-c1b0-45dc-81ab-a38827c116ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.447962] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4055cd95-69c1-45bb-a9b5-62285020cb17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.477033] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6eea6917-0ed4-447a-96f0-094c1832160a could not be found. [ 524.477033] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.477033] env[62066]: INFO nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.477033] env[62066]: DEBUG oslo.service.loopingcall [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.477033] env[62066]: DEBUG nova.compute.manager [-] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.477033] env[62066]: DEBUG nova.network.neutron [-] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.510552] env[62066]: DEBUG nova.network.neutron [-] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.664100] env[62066]: DEBUG nova.policy [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '340f5d349bf848919d32407ed28d9949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a079fb2263e4444880bd8f56fa0e28eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.735319] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.793599] env[62066]: ERROR nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 524.793599] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.793599] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.793599] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.793599] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.793599] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.793599] env[62066]: ERROR nova.compute.manager raise self.value [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.793599] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.793599] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.793599] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.794169] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.794169] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.794169] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 524.794169] env[62066]: ERROR nova.compute.manager [ 524.794169] env[62066]: Traceback (most recent call last): [ 524.794169] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.794169] env[62066]: listener.cb(fileno) [ 524.794169] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.794169] env[62066]: result = function(*args, **kwargs) [ 524.794169] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.794169] env[62066]: return func(*args, **kwargs) [ 524.794169] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.794169] env[62066]: raise e [ 524.794169] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.794169] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 524.794169] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.794169] env[62066]: created_port_ids = self._update_ports_for_instance( [ 524.794169] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.794169] env[62066]: with excutils.save_and_reraise_exception(): [ 524.794169] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.794169] env[62066]: self.force_reraise() [ 524.794169] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.794169] env[62066]: raise self.value [ 524.794169] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.794169] env[62066]: updated_port = self._update_port( [ 524.794169] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.794169] env[62066]: _ensure_no_port_binding_failure(port) [ 524.794169] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.794169] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.794985] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 524.794985] env[62066]: Removing descriptor: 18 [ 524.794985] env[62066]: ERROR nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Traceback (most recent call last): [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] yield resources [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.driver.spawn(context, instance, image_meta, [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.794985] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] vm_ref = self.build_virtual_machine(instance, [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] for vif in network_info: [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self._sync_wrapper(fn, *args, **kwargs) [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.wait() [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self[:] = self._gt.wait() [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self._exit_event.wait() [ 524.795408] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] result = hub.switch() [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self.greenlet.switch() [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] result = function(*args, **kwargs) [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return func(*args, **kwargs) [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise e [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] nwinfo = self.network_api.allocate_for_instance( [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.795840] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] created_port_ids = self._update_ports_for_instance( [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] with excutils.save_and_reraise_exception(): [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.force_reraise() [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise self.value [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] updated_port = self._update_port( [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] _ensure_no_port_binding_failure(port) [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.796260] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise exception.PortBindingFailed(port_id=port['id']) [ 524.796594] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 524.796594] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] [ 524.796594] env[62066]: INFO nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Terminating instance [ 524.797377] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquiring lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.797676] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquired lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.797952] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.925242] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.020133] env[62066]: DEBUG nova.network.neutron [-] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.105888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dde5442-7259-44d4-938b-1639f5cf0b84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.115406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a063d5c-e395-476e-bfb2-1a41a7a79685 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.155714] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d62bd89-0bb3-421d-ae42-c374cadef880 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.163986] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17d7007-8bbd-4d0d-b76a-405070cb0b0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.179077] env[62066]: DEBUG nova.compute.provider_tree [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.263662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.346271] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.463951] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.522232] env[62066]: INFO nova.compute.manager [-] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Took 1.04 seconds to deallocate network for instance. [ 525.526579] env[62066]: DEBUG nova.compute.claims [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.526786] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.682621] env[62066]: DEBUG nova.scheduler.client.report [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.795560] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Successfully created port: bb1e784e-e960-4404-b7ab-b423a47bdabe {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 525.953832] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.967689] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Releasing lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.969730] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.969730] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.969730] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06a88c48-b75c-4fa7-bc92-8f90b157d057 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.983341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c128d951-8b73-47e2-af9c-354f24956711 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.014427] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b78d0ac-2874-417a-ad2d-d4197d6fbe95 could not be found. [ 526.014660] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 526.014853] env[62066]: INFO nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Took 0.05 seconds to destroy the instance on the hypervisor. [ 526.015133] env[62066]: DEBUG oslo.service.loopingcall [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.015527] env[62066]: DEBUG nova.compute.manager [-] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.015618] env[62066]: DEBUG nova.network.neutron [-] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.029946] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.030219] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.030370] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.030544] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.030684] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.030822] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.031036] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.031188] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.031353] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.031511] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.031672] env[62066]: DEBUG nova.virt.hardware [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.032664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1c3ce3-4b3a-4df3-aa94-3b9feefc01ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.044519] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c502f28-16ad-44f5-89b4-fed04d0bf30c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.051298] env[62066]: DEBUG nova.network.neutron [-] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.065826] env[62066]: DEBUG nova.network.neutron [-] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.175795] env[62066]: DEBUG nova.compute.manager [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Received event network-changed-26bc6c6c-abcb-4f06-b02e-020492ead1c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.175927] env[62066]: DEBUG nova.compute.manager [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Refreshing instance network info cache due to event network-changed-26bc6c6c-abcb-4f06-b02e-020492ead1c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.176304] env[62066]: DEBUG oslo_concurrency.lockutils [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] Acquiring lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.176304] env[62066]: DEBUG oslo_concurrency.lockutils [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] Acquired lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.176469] env[62066]: DEBUG nova.network.neutron [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Refreshing network info cache for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 526.189813] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.190345] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.193530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.044s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.194770] env[62066]: INFO nova.compute.claims [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.316381] env[62066]: ERROR nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 526.316381] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.316381] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.316381] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.316381] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.316381] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.316381] env[62066]: ERROR nova.compute.manager raise self.value [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.316381] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.316381] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.316381] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.316850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.316850] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.316850] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 526.316850] env[62066]: ERROR nova.compute.manager [ 526.316850] env[62066]: Traceback (most recent call last): [ 526.316850] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.316850] env[62066]: listener.cb(fileno) [ 526.316850] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.316850] env[62066]: result = function(*args, **kwargs) [ 526.316850] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.316850] env[62066]: return func(*args, **kwargs) [ 526.316850] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.316850] env[62066]: raise e [ 526.316850] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.316850] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 526.316850] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.316850] env[62066]: created_port_ids = self._update_ports_for_instance( [ 526.316850] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.316850] env[62066]: with excutils.save_and_reraise_exception(): [ 526.316850] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.316850] env[62066]: self.force_reraise() [ 526.316850] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.316850] env[62066]: raise self.value [ 526.316850] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.316850] env[62066]: updated_port = self._update_port( [ 526.316850] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.316850] env[62066]: _ensure_no_port_binding_failure(port) [ 526.316850] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.316850] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.317637] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 526.317637] env[62066]: Removing descriptor: 17 [ 526.317637] env[62066]: ERROR nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Traceback (most recent call last): [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] yield resources [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.driver.spawn(context, instance, image_meta, [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.317637] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] vm_ref = self.build_virtual_machine(instance, [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] for vif in network_info: [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self._sync_wrapper(fn, *args, **kwargs) [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.wait() [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self[:] = self._gt.wait() [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self._exit_event.wait() [ 526.318025] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] result = hub.switch() [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self.greenlet.switch() [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] result = function(*args, **kwargs) [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return func(*args, **kwargs) [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise e [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] nwinfo = self.network_api.allocate_for_instance( [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.321055] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] created_port_ids = self._update_ports_for_instance( [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] with excutils.save_and_reraise_exception(): [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.force_reraise() [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise self.value [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] updated_port = self._update_port( [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] _ensure_no_port_binding_failure(port) [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.322046] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise exception.PortBindingFailed(port_id=port['id']) [ 526.323471] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 526.323471] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] [ 526.323471] env[62066]: INFO nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Terminating instance [ 526.323471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.323471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquired lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.323471] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 526.569579] env[62066]: INFO nova.compute.manager [-] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Took 0.55 seconds to deallocate network for instance. [ 526.571889] env[62066]: DEBUG nova.compute.claims [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.572125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.706499] env[62066]: DEBUG nova.compute.utils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 526.706499] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.706499] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 526.714473] env[62066]: DEBUG nova.network.neutron [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.807652] env[62066]: DEBUG nova.compute.manager [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Received event network-changed-4866db3b-9a37-48d8-9bb8-3127ce1f124e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.807652] env[62066]: DEBUG nova.compute.manager [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Refreshing instance network info cache due to event network-changed-4866db3b-9a37-48d8-9bb8-3127ce1f124e. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.807652] env[62066]: DEBUG oslo_concurrency.lockutils [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] Acquiring lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.807652] env[62066]: DEBUG oslo_concurrency.lockutils [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] Acquired lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.807652] env[62066]: DEBUG nova.network.neutron [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Refreshing network info cache for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 526.838226] env[62066]: DEBUG nova.policy [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79e76628ac91496d9c4959d25c609b79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75bba63736944f92ba5c09bd23f2407a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.861615] env[62066]: DEBUG nova.network.neutron [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.867671] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.054399] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.210226] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.338074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9a48ba-88bc-4894-8fc7-7af833d3e130 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.346757] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bfaf90-c601-48a3-9c9c-65bca4c52294 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.379604] env[62066]: DEBUG oslo_concurrency.lockutils [req-065a4c3d-49ef-4b2e-9c2f-8c1f8f7af59c req-56c8c43c-3ee6-4237-bf80-2a6b088c14d5 service nova] Releasing lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.380684] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1e8b57-0b72-4a0a-9720-d963c3fb2157 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.389200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d499b8c-220a-494f-9d3a-9475a0672042 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.402716] env[62066]: DEBUG nova.compute.provider_tree [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.453860] env[62066]: DEBUG nova.network.neutron [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.557031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Releasing lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.557482] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.557661] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.557958] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-135a73ef-a5b7-44e6-a8e8-641d79b6e612 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.578832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a267929-b24b-4e4c-8722-9905da713e5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.601977] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a233c561-06e9-49c8-8c3d-4211acfbdd4d could not be found. [ 527.602224] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.602408] env[62066]: INFO nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 527.602692] env[62066]: DEBUG oslo.service.loopingcall [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.602901] env[62066]: DEBUG nova.compute.manager [-] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.602995] env[62066]: DEBUG nova.network.neutron [-] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 527.632872] env[62066]: DEBUG nova.network.neutron [-] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.665993] env[62066]: DEBUG nova.network.neutron [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.907393] env[62066]: DEBUG nova.scheduler.client.report [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.947647] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Successfully created port: a4fa6a13-9e3a-4092-95a2-4871e18a5de7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.135321] env[62066]: DEBUG nova.network.neutron [-] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.168557] env[62066]: DEBUG oslo_concurrency.lockutils [req-dfb85c1e-77b1-411f-819a-a4219d6b7bd5 req-17c08e49-e376-4b8c-bba8-5769870bbc07 service nova] Releasing lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.225049] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.262804] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.263498] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.263498] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.263498] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.263692] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.263692] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.263854] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.264017] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.264194] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.264354] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.264580] env[62066]: DEBUG nova.virt.hardware [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.265434] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13d1f6f-b368-41d4-929f-ddfe31109e08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.275172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f77cab-8c92-4cee-9057-778eb5264a5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.412608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.219s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.414609] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 528.419194] env[62066]: ERROR nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 528.419194] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.419194] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.419194] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.419194] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.419194] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.419194] env[62066]: ERROR nova.compute.manager raise self.value [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.419194] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.419194] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.419194] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.419690] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.419690] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.419690] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 528.419690] env[62066]: ERROR nova.compute.manager [ 528.425037] env[62066]: Traceback (most recent call last): [ 528.425037] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.425037] env[62066]: listener.cb(fileno) [ 528.425037] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.425037] env[62066]: result = function(*args, **kwargs) [ 528.425037] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.425037] env[62066]: return func(*args, **kwargs) [ 528.425037] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.425037] env[62066]: raise e [ 528.425037] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.425037] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 528.425037] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.425037] env[62066]: created_port_ids = self._update_ports_for_instance( [ 528.425037] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.425037] env[62066]: with excutils.save_and_reraise_exception(): [ 528.425037] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.425037] env[62066]: self.force_reraise() [ 528.425037] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.425037] env[62066]: raise self.value [ 528.425037] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.425037] env[62066]: updated_port = self._update_port( [ 528.425037] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.425037] env[62066]: _ensure_no_port_binding_failure(port) [ 528.425037] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.425037] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.425037] env[62066]: nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 528.425037] env[62066]: Removing descriptor: 16 [ 528.425956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.157s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.425956] env[62066]: INFO nova.compute.claims [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.427689] env[62066]: ERROR nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] Traceback (most recent call last): [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] yield resources [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.driver.spawn(context, instance, image_meta, [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] vm_ref = self.build_virtual_machine(instance, [ 528.427689] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] for vif in network_info: [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self._sync_wrapper(fn, *args, **kwargs) [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.wait() [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self[:] = self._gt.wait() [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self._exit_event.wait() [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.428041] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] result = hub.switch() [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self.greenlet.switch() [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] result = function(*args, **kwargs) [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return func(*args, **kwargs) [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise e [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] nwinfo = self.network_api.allocate_for_instance( [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] created_port_ids = self._update_ports_for_instance( [ 528.428402] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] with excutils.save_and_reraise_exception(): [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.force_reraise() [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise self.value [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] updated_port = self._update_port( [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] _ensure_no_port_binding_failure(port) [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise exception.PortBindingFailed(port_id=port['id']) [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 528.428790] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] [ 528.429217] env[62066]: INFO nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Terminating instance [ 528.437106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.437307] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.437494] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 528.639609] env[62066]: INFO nova.compute.manager [-] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Took 1.04 seconds to deallocate network for instance. [ 528.642836] env[62066]: DEBUG nova.compute.claims [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.642956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.936269] env[62066]: DEBUG nova.compute.utils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 528.937966] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.938874] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.999754] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.079434] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Received event network-vif-deleted-26bc6c6c-abcb-4f06-b02e-020492ead1c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.079753] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Received event network-changed-30270b72-5e44-481c-9660-885918d4c9e4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.079862] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Refreshing instance network info cache due to event network-changed-30270b72-5e44-481c-9660-885918d4c9e4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.080106] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Acquiring lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.080252] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Acquired lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.080437] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Refreshing network info cache for port 30270b72-5e44-481c-9660-885918d4c9e4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 529.110941] env[62066]: DEBUG nova.policy [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '340f5d349bf848919d32407ed28d9949', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a079fb2263e4444880bd8f56fa0e28eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.408878] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.444027] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 529.588123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ff0812-0e59-49b6-9fa3-da14d7f4151f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.599561] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1432f90c-19cc-4f2a-ad97-a4ff0586785b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.633826] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.637279] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8844bb07-ab4b-4c6d-8c7f-9d029ecd8b7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.645657] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc2b77f-b3ae-462b-83fe-38ee4d6d9af5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.660435] env[62066]: DEBUG nova.compute.provider_tree [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.820692] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.913033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.913033] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.913033] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.913033] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e0e2ea1-70dd-45c1-917b-0bb7ea27d953 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.922791] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457ae1c7-f3e2-4733-97ba-6fded9c8e947 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.959010] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cdccd327-293e-4401-905c-4c16150e493a could not be found. [ 529.960441] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.960516] env[62066]: INFO nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 529.963871] env[62066]: DEBUG oslo.service.loopingcall [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.963871] env[62066]: DEBUG nova.compute.manager [-] [instance: cdccd327-293e-4401-905c-4c16150e493a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.963871] env[62066]: DEBUG nova.network.neutron [-] [instance: cdccd327-293e-4401-905c-4c16150e493a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.000936] env[62066]: DEBUG nova.network.neutron [-] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.098320] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Successfully created port: bcf3a749-b9a8-44cf-8abb-78c6a97659a8 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.164836] env[62066]: DEBUG nova.scheduler.client.report [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 530.328036] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Releasing lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.328553] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Received event network-vif-deleted-30270b72-5e44-481c-9660-885918d4c9e4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.328771] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Received event network-changed-bb1e784e-e960-4404-b7ab-b423a47bdabe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.328998] env[62066]: DEBUG nova.compute.manager [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Refreshing instance network info cache due to event network-changed-bb1e784e-e960-4404-b7ab-b423a47bdabe. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 530.329403] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Acquiring lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.329566] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Acquired lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.329873] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Refreshing network info cache for port bb1e784e-e960-4404-b7ab-b423a47bdabe {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.452404] env[62066]: DEBUG nova.compute.manager [req-bfca4150-44a0-4fa2-8f9c-7245afbddfc3 req-2b3247b0-1d16-4a7f-ba12-b7fce51cb3a9 service nova] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Received event network-vif-deleted-4866db3b-9a37-48d8-9bb8-3127ce1f124e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.465513] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.503972] env[62066]: DEBUG nova.network.neutron [-] [instance: cdccd327-293e-4401-905c-4c16150e493a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.510222] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.510474] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.510644] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.510852] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.511009] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.511092] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.511299] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.511494] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.511678] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.511842] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.512017] env[62066]: DEBUG nova.virt.hardware [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.512976] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4cb89e-6ece-43f2-ab6e-cebf0d65cb7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.523046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964c70cb-7888-4fe5-b4e4-18c3c074a58f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.669281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.669734] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.676142] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.149s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.854482] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.009158] env[62066]: INFO nova.compute.manager [-] [instance: cdccd327-293e-4401-905c-4c16150e493a] Took 1.05 seconds to deallocate network for instance. [ 531.012287] env[62066]: DEBUG nova.compute.claims [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.012475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.050740] env[62066]: DEBUG nova.network.neutron [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.118851] env[62066]: ERROR nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 531.118851] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.118851] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 531.118851] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 531.118851] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.118851] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.118851] env[62066]: ERROR nova.compute.manager raise self.value [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 531.118851] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 531.118851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.118851] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 531.119337] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.119337] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 531.119337] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 531.119337] env[62066]: ERROR nova.compute.manager [ 531.119337] env[62066]: Traceback (most recent call last): [ 531.119337] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 531.119337] env[62066]: listener.cb(fileno) [ 531.119337] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.119337] env[62066]: result = function(*args, **kwargs) [ 531.119337] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.119337] env[62066]: return func(*args, **kwargs) [ 531.119337] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 531.119337] env[62066]: raise e [ 531.119337] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.119337] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 531.119337] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 531.119337] env[62066]: created_port_ids = self._update_ports_for_instance( [ 531.119766] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 531.119766] env[62066]: with excutils.save_and_reraise_exception(): [ 531.119766] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.119766] env[62066]: self.force_reraise() [ 531.119766] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.119766] env[62066]: raise self.value [ 531.119766] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 531.119766] env[62066]: updated_port = self._update_port( [ 531.119766] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.119766] env[62066]: _ensure_no_port_binding_failure(port) [ 531.119766] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.119766] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 531.119766] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 531.119766] env[62066]: Removing descriptor: 18 [ 531.121199] env[62066]: ERROR nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Traceback (most recent call last): [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] yield resources [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.driver.spawn(context, instance, image_meta, [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] vm_ref = self.build_virtual_machine(instance, [ 531.121199] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] vif_infos = vmwarevif.get_vif_info(self._session, [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] for vif in network_info: [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self._sync_wrapper(fn, *args, **kwargs) [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.wait() [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self[:] = self._gt.wait() [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self._exit_event.wait() [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 531.121565] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] result = hub.switch() [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self.greenlet.switch() [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] result = function(*args, **kwargs) [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return func(*args, **kwargs) [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise e [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] nwinfo = self.network_api.allocate_for_instance( [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] created_port_ids = self._update_ports_for_instance( [ 531.121891] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] with excutils.save_and_reraise_exception(): [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.force_reraise() [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise self.value [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] updated_port = self._update_port( [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] _ensure_no_port_binding_failure(port) [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise exception.PortBindingFailed(port_id=port['id']) [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 531.122266] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] [ 531.122667] env[62066]: INFO nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Terminating instance [ 531.123540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquiring lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.123700] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquired lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.123906] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 531.187246] env[62066]: DEBUG nova.compute.utils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 531.192212] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.192384] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 531.241695] env[62066]: DEBUG nova.policy [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b786921efb3540828b1079769c40059d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd06323d3542b4e5c9d61f34173da6895', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.412067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d297a63e-bbda-4c05-9edc-ad916a422942 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.421072] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5458003f-b408-4060-ba96-30494fcbf4cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.457953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402a06a7-75c2-44d5-915b-e9d81ca9f2df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.466225] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8610d8fc-fdd9-4c4f-9b4c-b64f1406ba62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.481150] env[62066]: DEBUG nova.compute.provider_tree [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.557075] env[62066]: DEBUG oslo_concurrency.lockutils [req-106150a7-8c41-41fb-9cb5-9f8770a0d099 req-ce7fcd2f-9834-4192-a53e-eccca76c3e8c service nova] Releasing lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.652084] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.694756] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.727862] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Successfully created port: 98d36abe-ea03-453f-852d-4b4723e0ab51 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.778138] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.987179] env[62066]: DEBUG nova.scheduler.client.report [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.269584] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.270994] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.270994] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 532.270994] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 532.282297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Releasing lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.282492] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.282934] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.282934] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acace4cf-9c18-4f40-a5a8-53907dc5fcbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.294525] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f342021f-8291-4130-af97-f61bf860cae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.320214] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bfa00759-94d1-47c5-9b43-53f093ef8fdc could not be found. [ 532.320422] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.320646] env[62066]: INFO nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 532.320898] env[62066]: DEBUG oslo.service.loopingcall [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.321122] env[62066]: DEBUG nova.compute.manager [-] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.321218] env[62066]: DEBUG nova.network.neutron [-] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 532.475068] env[62066]: DEBUG nova.network.neutron [-] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.492426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.816s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.493014] env[62066]: ERROR nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Traceback (most recent call last): [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.driver.spawn(context, instance, image_meta, [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] vm_ref = self.build_virtual_machine(instance, [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.493014] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] for vif in network_info: [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self._sync_wrapper(fn, *args, **kwargs) [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.wait() [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self[:] = self._gt.wait() [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self._exit_event.wait() [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] result = hub.switch() [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.493343] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return self.greenlet.switch() [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] result = function(*args, **kwargs) [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] return func(*args, **kwargs) [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise e [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] nwinfo = self.network_api.allocate_for_instance( [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] created_port_ids = self._update_ports_for_instance( [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] with excutils.save_and_reraise_exception(): [ 532.493659] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] self.force_reraise() [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise self.value [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] updated_port = self._update_port( [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] _ensure_no_port_binding_failure(port) [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] raise exception.PortBindingFailed(port_id=port['id']) [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] nova.exception.PortBindingFailed: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. [ 532.493980] env[62066]: ERROR nova.compute.manager [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] [ 532.495507] env[62066]: DEBUG nova.compute.utils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 532.496976] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.925s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.505375] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Build of instance 6eea6917-0ed4-447a-96f0-094c1832160a was re-scheduled: Binding failed for port 26bc6c6c-abcb-4f06-b02e-020492ead1c7, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 532.505670] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 532.505906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.506065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.506311] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 532.710308] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.741775] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.741775] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.741775] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.741933] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.742580] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.743082] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.743510] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.743999] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.744408] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.744946] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.745788] env[62066]: DEBUG nova.virt.hardware [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.747358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d4c687-8897-45ce-a32a-7751a31609c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.761376] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c7f08f-bb4a-4e2e-90fd-25bf3c9ce521 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.784766] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.784994] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.785198] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: cdccd327-293e-4401-905c-4c16150e493a] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.785625] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.785625] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.785625] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 532.785836] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 532.786352] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.786656] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.786919] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.787195] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.787452] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.787698] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_power_states {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.977844] env[62066]: DEBUG nova.network.neutron [-] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.033692] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.112847] env[62066]: ERROR nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 533.112847] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.112847] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.112847] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.112847] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.112847] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.112847] env[62066]: ERROR nova.compute.manager raise self.value [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.112847] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.112847] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.112847] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.113359] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.113359] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.113359] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 533.113359] env[62066]: ERROR nova.compute.manager [ 533.113359] env[62066]: Traceback (most recent call last): [ 533.113359] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.113359] env[62066]: listener.cb(fileno) [ 533.113359] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.113359] env[62066]: result = function(*args, **kwargs) [ 533.113359] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.113359] env[62066]: return func(*args, **kwargs) [ 533.113359] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.113359] env[62066]: raise e [ 533.113359] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.113359] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 533.113359] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.113359] env[62066]: created_port_ids = self._update_ports_for_instance( [ 533.113359] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.113359] env[62066]: with excutils.save_and_reraise_exception(): [ 533.113359] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.113359] env[62066]: self.force_reraise() [ 533.113359] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.113359] env[62066]: raise self.value [ 533.113359] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.113359] env[62066]: updated_port = self._update_port( [ 533.113359] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.113359] env[62066]: _ensure_no_port_binding_failure(port) [ 533.113359] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.113359] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.114497] env[62066]: nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 533.114497] env[62066]: Removing descriptor: 16 [ 533.114497] env[62066]: ERROR nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Traceback (most recent call last): [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] yield resources [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.driver.spawn(context, instance, image_meta, [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.114497] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] vm_ref = self.build_virtual_machine(instance, [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] for vif in network_info: [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self._sync_wrapper(fn, *args, **kwargs) [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.wait() [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self[:] = self._gt.wait() [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self._exit_event.wait() [ 533.114891] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] result = hub.switch() [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self.greenlet.switch() [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] result = function(*args, **kwargs) [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return func(*args, **kwargs) [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise e [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] nwinfo = self.network_api.allocate_for_instance( [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.115250] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] created_port_ids = self._update_ports_for_instance( [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] with excutils.save_and_reraise_exception(): [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.force_reraise() [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise self.value [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] updated_port = self._update_port( [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] _ensure_no_port_binding_failure(port) [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.115570] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise exception.PortBindingFailed(port_id=port['id']) [ 533.115928] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 533.115928] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] [ 533.115928] env[62066]: INFO nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Terminating instance [ 533.118651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.118651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.118651] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.125326] env[62066]: DEBUG nova.compute.manager [req-d4ecbefb-032f-49f7-b155-785dea9b91cf req-4982f0ad-52cb-47e2-9506-681fd21a0615 service nova] [instance: cdccd327-293e-4401-905c-4c16150e493a] Received event network-vif-deleted-bb1e784e-e960-4404-b7ab-b423a47bdabe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.128080] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce46c5fb-37ce-49aa-86f7-6b29430ea1b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.138177] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefa7bb5-ff1d-4ab1-86da-e0408d575d59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.173616] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f423120-68a3-4027-8a6e-6c0bd5524bf2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.181074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbea9df-20dc-45a7-bd93-2caceaa0676f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.195450] env[62066]: DEBUG nova.compute.provider_tree [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.204922] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.295898] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Getting list of instances from cluster (obj){ [ 533.295898] env[62066]: value = "domain-c8" [ 533.295898] env[62066]: _type = "ClusterComputeResource" [ 533.295898] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 533.295898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e291b033-bc77-4fed-b34f-18a2cd970378 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.302971] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Got total of 0 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 533.303165] env[62066]: WARNING nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] While synchronizing instance power states, found 5 instances in the database and 0 instances on the hypervisor. [ 533.303310] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid a233c561-06e9-49c8-8c3d-4211acfbdd4d {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 533.303506] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid cdccd327-293e-4401-905c-4c16150e493a {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 533.303663] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid bfa00759-94d1-47c5-9b43-53f093ef8fdc {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 533.303810] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid ed41309e-0253-4bbf-acd5-e1030ce2aed0 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 533.303963] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid 6e2496af-4d60-4315-9816-d0456458bc13 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 533.304478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.304740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "cdccd327-293e-4401-905c-4c16150e493a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.304949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.305170] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.305366] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "6e2496af-4d60-4315-9816-d0456458bc13" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.305897] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.306168] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 533.306341] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.418821] env[62066]: ERROR nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 533.418821] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.418821] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.418821] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.418821] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.418821] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.418821] env[62066]: ERROR nova.compute.manager raise self.value [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.418821] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.418821] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.418821] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.419895] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.419895] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.419895] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 533.419895] env[62066]: ERROR nova.compute.manager [ 533.419895] env[62066]: Traceback (most recent call last): [ 533.419895] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.419895] env[62066]: listener.cb(fileno) [ 533.419895] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.419895] env[62066]: result = function(*args, **kwargs) [ 533.419895] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.419895] env[62066]: return func(*args, **kwargs) [ 533.419895] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.419895] env[62066]: raise e [ 533.419895] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.419895] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 533.419895] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.419895] env[62066]: created_port_ids = self._update_ports_for_instance( [ 533.419895] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.419895] env[62066]: with excutils.save_and_reraise_exception(): [ 533.419895] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.419895] env[62066]: self.force_reraise() [ 533.419895] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.419895] env[62066]: raise self.value [ 533.419895] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.419895] env[62066]: updated_port = self._update_port( [ 533.419895] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.419895] env[62066]: _ensure_no_port_binding_failure(port) [ 533.419895] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.419895] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.421602] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 533.421602] env[62066]: Removing descriptor: 18 [ 533.421602] env[62066]: ERROR nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Traceback (most recent call last): [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] yield resources [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.driver.spawn(context, instance, image_meta, [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.421602] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] vm_ref = self.build_virtual_machine(instance, [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] for vif in network_info: [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self._sync_wrapper(fn, *args, **kwargs) [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.wait() [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self[:] = self._gt.wait() [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self._exit_event.wait() [ 533.422161] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] result = hub.switch() [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self.greenlet.switch() [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] result = function(*args, **kwargs) [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return func(*args, **kwargs) [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise e [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] nwinfo = self.network_api.allocate_for_instance( [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.422645] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] created_port_ids = self._update_ports_for_instance( [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] with excutils.save_and_reraise_exception(): [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.force_reraise() [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise self.value [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] updated_port = self._update_port( [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] _ensure_no_port_binding_failure(port) [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.422970] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise exception.PortBindingFailed(port_id=port['id']) [ 533.423295] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 533.423295] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] [ 533.423295] env[62066]: INFO nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Terminating instance [ 533.423295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.423295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquired lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.423295] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.483324] env[62066]: INFO nova.compute.manager [-] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Took 1.16 seconds to deallocate network for instance. [ 533.487164] env[62066]: DEBUG nova.compute.claims [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.487290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.655464] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.698145] env[62066]: DEBUG nova.scheduler.client.report [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.709563] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-6eea6917-0ed4-447a-96f0-094c1832160a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.709563] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 533.709563] env[62066]: DEBUG nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.709563] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.749197] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.783401] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquiring lock "a87ad1d1-9613-4144-9390-d5e3da3a2806" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.783401] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "a87ad1d1-9613-4144-9390-d5e3da3a2806" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.810665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.851403] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.957226] env[62066]: DEBUG nova.compute.manager [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Received event network-changed-a4fa6a13-9e3a-4092-95a2-4871e18a5de7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.957226] env[62066]: DEBUG nova.compute.manager [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Refreshing instance network info cache due to event network-changed-a4fa6a13-9e3a-4092-95a2-4871e18a5de7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.957226] env[62066]: DEBUG oslo_concurrency.lockutils [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] Acquiring lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.957226] env[62066]: DEBUG oslo_concurrency.lockutils [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] Acquired lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.957226] env[62066]: DEBUG nova.network.neutron [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Refreshing network info cache for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 533.959509] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.035219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquiring lock "7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.035219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.178084] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.203910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.707s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.204855] env[62066]: ERROR nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Traceback (most recent call last): [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.driver.spawn(context, instance, image_meta, [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] vm_ref = self.build_virtual_machine(instance, [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.204855] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] for vif in network_info: [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self._sync_wrapper(fn, *args, **kwargs) [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.wait() [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self[:] = self._gt.wait() [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self._exit_event.wait() [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] result = hub.switch() [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.205285] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return self.greenlet.switch() [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] result = function(*args, **kwargs) [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] return func(*args, **kwargs) [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise e [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] nwinfo = self.network_api.allocate_for_instance( [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] created_port_ids = self._update_ports_for_instance( [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] with excutils.save_and_reraise_exception(): [ 534.205646] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] self.force_reraise() [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise self.value [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] updated_port = self._update_port( [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] _ensure_no_port_binding_failure(port) [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] raise exception.PortBindingFailed(port_id=port['id']) [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] nova.exception.PortBindingFailed: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. [ 534.205961] env[62066]: ERROR nova.compute.manager [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] [ 534.206297] env[62066]: DEBUG nova.compute.utils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 534.210037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.564s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.210630] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Build of instance 5b78d0ac-2874-417a-ad2d-d4197d6fbe95 was re-scheduled: Binding failed for port 4866db3b-9a37-48d8-9bb8-3127ce1f124e, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 534.211101] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 534.211333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquiring lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.211488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Acquired lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.211813] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.253328] env[62066]: DEBUG nova.network.neutron [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.286519] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.353633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.353959] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.354182] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.354712] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fec828bf-57c6-480a-a363-b117196c185f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.363859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38695c01-740c-4f12-9965-daf18973b824 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.385693] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed41309e-0253-4bbf-acd5-e1030ce2aed0 could not be found. [ 534.385912] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.386107] env[62066]: INFO nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 534.386337] env[62066]: DEBUG oslo.service.loopingcall [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.386624] env[62066]: DEBUG nova.compute.manager [-] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.386624] env[62066]: DEBUG nova.network.neutron [-] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.417161] env[62066]: DEBUG nova.network.neutron [-] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.506627] env[62066]: DEBUG nova.network.neutron [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.537845] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.680700] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Releasing lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.681163] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.681357] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.681654] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e859ced-950b-42ee-acc5-6f3e9108453c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.690905] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f41962c-8d41-43e8-96a1-8b9875409a61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.703259] env[62066]: DEBUG nova.network.neutron [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.720773] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e2496af-4d60-4315-9816-d0456458bc13 could not be found. [ 534.720979] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.721382] env[62066]: INFO nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Took 0.04 seconds to destroy the instance on the hypervisor. [ 534.721671] env[62066]: DEBUG oslo.service.loopingcall [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.723296] env[62066]: DEBUG nova.compute.manager [-] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.723449] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.756390] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.758522] env[62066]: INFO nova.compute.manager [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: 6eea6917-0ed4-447a-96f0-094c1832160a] Took 1.05 seconds to deallocate network for instance. [ 534.764056] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.815719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.847148] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.873150] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafcef4d-74fb-4145-883a-6fe9915ea866 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.881727] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47950a21-38e7-4bf5-9db4-3baa6fd4abbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.922321] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e1e764-3d24-45c3-996b-56f6e170a216 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.923423] env[62066]: DEBUG nova.network.neutron [-] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.930477] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446e8cec-2a54-45ea-851d-d8ee151377d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.951406] env[62066]: DEBUG nova.compute.provider_tree [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.061952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.206450] env[62066]: DEBUG oslo_concurrency.lockutils [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] Releasing lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.206542] env[62066]: DEBUG nova.compute.manager [req-1603ac54-2eb8-48b7-9614-0dd4a4a38b74 req-7ba8400e-85fa-4823-9f6e-fd16e88248e7 service nova] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Received event network-vif-deleted-a4fa6a13-9e3a-4092-95a2-4871e18a5de7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.269803] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.349830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Releasing lock "refresh_cache-5b78d0ac-2874-417a-ad2d-d4197d6fbe95" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.350229] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 535.350340] env[62066]: DEBUG nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.350558] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.375131] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.426563] env[62066]: INFO nova.compute.manager [-] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Took 1.04 seconds to deallocate network for instance. [ 535.428841] env[62066]: DEBUG nova.compute.claims [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.429071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.454650] env[62066]: DEBUG nova.scheduler.client.report [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.774754] env[62066]: INFO nova.compute.manager [-] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Took 1.05 seconds to deallocate network for instance. [ 535.777499] env[62066]: DEBUG nova.compute.claims [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.777683] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.800558] env[62066]: INFO nova.scheduler.client.report [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Deleted allocations for instance 6eea6917-0ed4-447a-96f0-094c1832160a [ 535.878292] env[62066]: DEBUG nova.network.neutron [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.961761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.754s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.962140] env[62066]: ERROR nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Traceback (most recent call last): [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.driver.spawn(context, instance, image_meta, [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] vm_ref = self.build_virtual_machine(instance, [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.962140] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] for vif in network_info: [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self._sync_wrapper(fn, *args, **kwargs) [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.wait() [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self[:] = self._gt.wait() [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self._exit_event.wait() [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] result = hub.switch() [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.962563] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return self.greenlet.switch() [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] result = function(*args, **kwargs) [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] return func(*args, **kwargs) [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise e [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] nwinfo = self.network_api.allocate_for_instance( [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] created_port_ids = self._update_ports_for_instance( [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] with excutils.save_and_reraise_exception(): [ 535.963023] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] self.force_reraise() [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise self.value [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] updated_port = self._update_port( [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] _ensure_no_port_binding_failure(port) [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] raise exception.PortBindingFailed(port_id=port['id']) [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] nova.exception.PortBindingFailed: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. [ 535.963416] env[62066]: ERROR nova.compute.manager [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] [ 535.963749] env[62066]: DEBUG nova.compute.utils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 535.964405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.952s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.967512] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Build of instance a233c561-06e9-49c8-8c3d-4211acfbdd4d was re-scheduled: Binding failed for port 30270b72-5e44-481c-9660-885918d4c9e4, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.967778] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.969803] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.970164] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquired lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.970404] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.311073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a48b79df-058e-48c3-9087-50578a8c48e9 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "6eea6917-0ed4-447a-96f0-094c1832160a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.419s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.382314] env[62066]: INFO nova.compute.manager [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] [instance: 5b78d0ac-2874-417a-ad2d-d4197d6fbe95] Took 1.03 seconds to deallocate network for instance. [ 536.553229] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.639903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481dd046-2f75-45ce-9eb6-ffc18ba60646 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.649804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f206d89-12f1-4a93-8de7-2e1d7a07c9f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.684436] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6981c621-c5d1-4603-bc32-a33199ed6211 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.692804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b4e0db-944f-4c84-9862-ba3a4a5fa29e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.708384] env[62066]: DEBUG nova.compute.provider_tree [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.805090] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.211449] env[62066]: DEBUG nova.scheduler.client.report [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.307737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Releasing lock "refresh_cache-a233c561-06e9-49c8-8c3d-4211acfbdd4d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.308010] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 537.308164] env[62066]: DEBUG nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.308371] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.363805] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.419154] env[62066]: INFO nova.scheduler.client.report [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Deleted allocations for instance 5b78d0ac-2874-417a-ad2d-d4197d6fbe95 [ 537.716524] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.752s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.717145] env[62066]: ERROR nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] Traceback (most recent call last): [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.driver.spawn(context, instance, image_meta, [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] vm_ref = self.build_virtual_machine(instance, [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.717145] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] for vif in network_info: [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self._sync_wrapper(fn, *args, **kwargs) [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.wait() [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self[:] = self._gt.wait() [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self._exit_event.wait() [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] result = hub.switch() [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.717484] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return self.greenlet.switch() [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] result = function(*args, **kwargs) [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] return func(*args, **kwargs) [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise e [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] nwinfo = self.network_api.allocate_for_instance( [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] created_port_ids = self._update_ports_for_instance( [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] with excutils.save_and_reraise_exception(): [ 537.717869] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] self.force_reraise() [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise self.value [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] updated_port = self._update_port( [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] _ensure_no_port_binding_failure(port) [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] raise exception.PortBindingFailed(port_id=port['id']) [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] nova.exception.PortBindingFailed: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. [ 537.719132] env[62066]: ERROR nova.compute.manager [instance: cdccd327-293e-4401-905c-4c16150e493a] [ 537.719469] env[62066]: DEBUG nova.compute.utils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 537.719469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.232s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.721963] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Build of instance cdccd327-293e-4401-905c-4c16150e493a was re-scheduled: Binding failed for port bb1e784e-e960-4404-b7ab-b423a47bdabe, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 537.722905] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 537.723187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.723376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.723591] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 537.739149] env[62066]: DEBUG nova.compute.manager [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Received event network-changed-bcf3a749-b9a8-44cf-8abb-78c6a97659a8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.739362] env[62066]: DEBUG nova.compute.manager [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Refreshing instance network info cache due to event network-changed-bcf3a749-b9a8-44cf-8abb-78c6a97659a8. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 537.739581] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] Acquiring lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.739720] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] Acquired lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.739872] env[62066]: DEBUG nova.network.neutron [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Refreshing network info cache for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 537.867793] env[62066]: DEBUG nova.network.neutron [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.923386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "fc176994-95b2-44a3-becf-6c224476851a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.923626] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "fc176994-95b2-44a3-becf-6c224476851a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.930098] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39d9dbd-72e1-4db8-8533-68184d46db71 tempest-ServerDiagnosticsNegativeTest-626980048 tempest-ServerDiagnosticsNegativeTest-626980048-project-member] Lock "5b78d0ac-2874-417a-ad2d-d4197d6fbe95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.019s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.013015] env[62066]: DEBUG nova.compute.manager [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Received event network-changed-98d36abe-ea03-453f-852d-4b4723e0ab51 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.013284] env[62066]: DEBUG nova.compute.manager [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Refreshing instance network info cache due to event network-changed-98d36abe-ea03-453f-852d-4b4723e0ab51. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.013584] env[62066]: DEBUG oslo_concurrency.lockutils [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] Acquiring lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.013738] env[62066]: DEBUG oslo_concurrency.lockutils [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] Acquired lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.013947] env[62066]: DEBUG nova.network.neutron [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Refreshing network info cache for port 98d36abe-ea03-453f-852d-4b4723e0ab51 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.162323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "fa73770f-784a-44c3-ba65-e7176180760b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.162570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.292447] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.298093] env[62066]: DEBUG nova.network.neutron [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.371448] env[62066]: INFO nova.compute.manager [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] Took 1.06 seconds to deallocate network for instance. [ 538.409397] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253979d1-8c5c-46a2-8c06-692079aad887 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.417751] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c2c8e0-ce38-483f-87d0-68bac79beceb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.426713] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.459649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732e1979-6bf4-4033-8c1b-63464d8dcd30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.468197] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b3e6e4-c6fc-4992-bb10-fa89a2d4383c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.486689] env[62066]: DEBUG nova.compute.provider_tree [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.566782] env[62066]: DEBUG nova.network.neutron [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.667139] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.811203] env[62066]: DEBUG nova.network.neutron [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.813304] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.857314] env[62066]: DEBUG nova.network.neutron [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.984083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.993940] env[62066]: DEBUG nova.scheduler.client.report [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.197496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.317600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-cdccd327-293e-4401-905c-4c16150e493a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.317872] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 539.318320] env[62066]: DEBUG nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.318578] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.321370] env[62066]: DEBUG oslo_concurrency.lockutils [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] Releasing lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.322290] env[62066]: DEBUG nova.compute.manager [req-ed09d250-a51c-47b3-909b-21f67c4c565a req-45e44e20-b46a-4e9a-8cc4-6baf6d290b4f service nova] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Received event network-vif-deleted-98d36abe-ea03-453f-852d-4b4723e0ab51 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.361731] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] Releasing lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.361731] env[62066]: DEBUG nova.compute.manager [req-d8e916e2-8f1e-4bf8-bb6c-e245ba63d155 req-7317f70d-816d-4b7c-bfd8-6591ee8476b1 service nova] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Received event network-vif-deleted-bcf3a749-b9a8-44cf-8abb-78c6a97659a8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.371322] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.406588] env[62066]: INFO nova.scheduler.client.report [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Deleted allocations for instance a233c561-06e9-49c8-8c3d-4211acfbdd4d [ 539.501850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.782s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.503473] env[62066]: ERROR nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Traceback (most recent call last): [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.driver.spawn(context, instance, image_meta, [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] vm_ref = self.build_virtual_machine(instance, [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.503473] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] for vif in network_info: [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self._sync_wrapper(fn, *args, **kwargs) [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.wait() [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self[:] = self._gt.wait() [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self._exit_event.wait() [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] result = hub.switch() [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.503853] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return self.greenlet.switch() [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] result = function(*args, **kwargs) [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] return func(*args, **kwargs) [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise e [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] nwinfo = self.network_api.allocate_for_instance( [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] created_port_ids = self._update_ports_for_instance( [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] with excutils.save_and_reraise_exception(): [ 539.504232] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] self.force_reraise() [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise self.value [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] updated_port = self._update_port( [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] _ensure_no_port_binding_failure(port) [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] raise exception.PortBindingFailed(port_id=port['id']) [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] nova.exception.PortBindingFailed: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. [ 539.504580] env[62066]: ERROR nova.compute.manager [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] [ 539.508202] env[62066]: DEBUG nova.compute.utils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 539.509482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.699s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.509651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.509804] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 539.510100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.695s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.513301] env[62066]: INFO nova.compute.claims [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.523489] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Build of instance bfa00759-94d1-47c5-9b43-53f093ef8fdc was re-scheduled: Binding failed for port a4fa6a13-9e3a-4092-95a2-4871e18a5de7, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 539.523489] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 539.523489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquiring lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.523489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Acquired lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.523648] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.525244] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d88fb2-1214-4512-af48-a7bfd7a8e031 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.537503] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bbdd18-b724-4cd5-af99-de5799831abe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.556812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3b4c1e-e49b-4b5b-9e59-f703351ce453 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.563956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fe9894-3221-4624-9be2-df073e1bb11a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.604682] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181515MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 539.604832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.877921] env[62066]: DEBUG nova.network.neutron [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.921911] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e91209e-e5c8-416a-ba4e-cc92677dea0f tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.021s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.922088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.618s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.922298] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a233c561-06e9-49c8-8c3d-4211acfbdd4d] During sync_power_state the instance has a pending task (spawning). Skip. [ 539.922330] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "a233c561-06e9-49c8-8c3d-4211acfbdd4d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.058688] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.183459] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.323349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquiring lock "212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.323349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.385390] env[62066]: INFO nova.compute.manager [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: cdccd327-293e-4401-905c-4c16150e493a] Took 1.07 seconds to deallocate network for instance. [ 540.687012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Releasing lock "refresh_cache-bfa00759-94d1-47c5-9b43-53f093ef8fdc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.687012] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 540.687012] env[62066]: DEBUG nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.687012] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.711720] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448c10e2-d85f-425c-80ff-cc146b216701 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.715480] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.721835] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f29505-925e-46b1-8de6-0d3755dcb914 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.756848] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abe9d52-0470-45c6-814c-935a2031c325 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.767186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f96a2f-d57b-4b41-8efe-61c95b1a1e2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.780199] env[62066]: DEBUG nova.compute.provider_tree [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.825724] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.219717] env[62066]: DEBUG nova.network.neutron [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.283811] env[62066]: DEBUG nova.scheduler.client.report [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.366147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.426381] env[62066]: INFO nova.scheduler.client.report [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Deleted allocations for instance cdccd327-293e-4401-905c-4c16150e493a [ 541.725799] env[62066]: INFO nova.compute.manager [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] Took 1.04 seconds to deallocate network for instance. [ 541.793044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.793645] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.796745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.735s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.798573] env[62066]: INFO nova.compute.claims [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.941681] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8c2961a3-366c-4d08-9121-55a0cf4edeae tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "cdccd327-293e-4401-905c-4c16150e493a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.286s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.942055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "cdccd327-293e-4401-905c-4c16150e493a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.637s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.942055] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: cdccd327-293e-4401-905c-4c16150e493a] During sync_power_state the instance has a pending task (spawning). Skip. [ 541.942248] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "cdccd327-293e-4401-905c-4c16150e493a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.303795] env[62066]: DEBUG nova.compute.utils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 542.310453] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.310552] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.425810] env[62066]: DEBUG nova.policy [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b2607f42901458190da81da955505e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50569fb511d34966a20666c4de795019', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.721674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquiring lock "bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.724364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.769158] env[62066]: INFO nova.scheduler.client.report [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Deleted allocations for instance bfa00759-94d1-47c5-9b43-53f093ef8fdc [ 542.821289] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.017598] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dfb06f-23fc-473d-ab0a-4a3545432874 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.029104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd91f2b-3b3f-4ed7-9e40-44b9f5a11b6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.072851] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6209c3db-7701-4f9f-b51a-841f91a0e82b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.080912] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfc2a1f-6380-4d63-b321-5a2e426ae30f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.096379] env[62066]: DEBUG nova.compute.provider_tree [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.229415] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.283019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-798edb17-3a6c-4c15-a492-7803d1bafd64 tempest-ImagesOneServerNegativeTestJSON-335748635 tempest-ImagesOneServerNegativeTestJSON-335748635-project-member] Lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.660s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.283019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.977s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.283019] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: bfa00759-94d1-47c5-9b43-53f093ef8fdc] During sync_power_state the instance has a pending task (spawning). Skip. [ 543.283019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "bfa00759-94d1-47c5-9b43-53f093ef8fdc" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.602819] env[62066]: DEBUG nova.scheduler.client.report [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.672715] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Successfully created port: 225dc014-1d6f-42f8-a368-c3757c9165b1 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.759740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.830378] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.866500] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.866746] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.866901] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.868032] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.868032] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.868032] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.868219] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.868255] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.868415] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.869021] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.869021] env[62066]: DEBUG nova.virt.hardware [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.870358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51989c99-6758-4e61-b047-4226bcb248fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.878863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d32ee27-387f-4b77-adc6-d51062e836fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.110742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.110742] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 544.112897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.684s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.623475] env[62066]: DEBUG nova.compute.utils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 544.628096] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.628242] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 544.806418] env[62066]: DEBUG nova.policy [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb713a64049248d386b4bce2d7c1b61a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09858aba0a28406d8e0520be1b36dfa9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.813116] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2ff128-fa0e-4d6d-9b0e-f594e6ccf22f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.826358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1606f2-df89-4913-bcfc-c3462125104a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.864753] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f228d7d4-56c9-4b67-b734-9251ca634a1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.873662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea59fb63-9d95-4a17-8386-4f1d8cb5483f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.899989] env[62066]: DEBUG nova.compute.provider_tree [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.940887] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquiring lock "92ca97dd-cd7c-4612-85d9-a05f5a785d6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.941368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "92ca97dd-cd7c-4612-85d9-a05f5a785d6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.129854] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.403255] env[62066]: DEBUG nova.scheduler.client.report [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.443122] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.443350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.444688] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.918676] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.804s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.918676] env[62066]: ERROR nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Traceback (most recent call last): [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.driver.spawn(context, instance, image_meta, [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.918676] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] vm_ref = self.build_virtual_machine(instance, [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] for vif in network_info: [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self._sync_wrapper(fn, *args, **kwargs) [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.wait() [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self[:] = self._gt.wait() [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self._exit_event.wait() [ 545.919236] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] result = hub.switch() [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return self.greenlet.switch() [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] result = function(*args, **kwargs) [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] return func(*args, **kwargs) [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise e [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] nwinfo = self.network_api.allocate_for_instance( [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.919590] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] created_port_ids = self._update_ports_for_instance( [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] with excutils.save_and_reraise_exception(): [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] self.force_reraise() [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise self.value [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] updated_port = self._update_port( [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] _ensure_no_port_binding_failure(port) [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.919923] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] raise exception.PortBindingFailed(port_id=port['id']) [ 545.920337] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] nova.exception.PortBindingFailed: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. [ 545.920337] env[62066]: ERROR nova.compute.manager [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] [ 545.920337] env[62066]: DEBUG nova.compute.utils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 545.922163] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Build of instance ed41309e-0253-4bbf-acd5-e1030ce2aed0 was re-scheduled: Binding failed for port bcf3a749-b9a8-44cf-8abb-78c6a97659a8, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.923250] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.925366] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquiring lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.925366] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Acquired lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.925366] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.925649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.148s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.951683] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.987491] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.123103] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Successfully created port: 1a63f849-d8b2-4293-93c6-9559584e6ed3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.142686] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 546.178361] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.178975] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.179316] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.179748] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.180178] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.183106] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.183106] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.183106] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.183106] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.183106] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.183306] env[62066]: DEBUG nova.virt.hardware [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.183306] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ff1a5c-068e-440f-9a7e-e755665f3743 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.192403] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e341599d-8a0d-4ef0-9361-1beb360a0473 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.472402] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.515400] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.593624] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719d600a-b3aa-4d2c-ac36-50cfb73760f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.604928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac25c13-9de8-4710-ac1d-c6d6c49435b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.648777] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb81bbd-041b-4245-b62a-c1c1a6105c72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.657600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d5b075-4045-4f57-bf2e-62c281e068b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.673241] env[62066]: DEBUG nova.compute.provider_tree [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.889279] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.180542] env[62066]: DEBUG nova.scheduler.client.report [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.392418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Releasing lock "refresh_cache-ed41309e-0253-4bbf-acd5-e1030ce2aed0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.393874] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 547.393874] env[62066]: DEBUG nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.393982] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.425396] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.592047] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquiring lock "b3d0de63-fe22-4bc7-9602-509054dca586" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.593810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "b3d0de63-fe22-4bc7-9602-509054dca586" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.683293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "cf3dc6c9-d7bf-4319-9420-e3634b9fca11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.683567] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "cf3dc6c9-d7bf-4319-9420-e3634b9fca11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.687814] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.762s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.688514] env[62066]: ERROR nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Traceback (most recent call last): [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.driver.spawn(context, instance, image_meta, [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] vm_ref = self.build_virtual_machine(instance, [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.688514] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] for vif in network_info: [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self._sync_wrapper(fn, *args, **kwargs) [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.wait() [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self[:] = self._gt.wait() [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self._exit_event.wait() [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] result = hub.switch() [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.688863] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return self.greenlet.switch() [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] result = function(*args, **kwargs) [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] return func(*args, **kwargs) [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise e [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] nwinfo = self.network_api.allocate_for_instance( [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] created_port_ids = self._update_ports_for_instance( [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] with excutils.save_and_reraise_exception(): [ 547.689387] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] self.force_reraise() [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise self.value [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] updated_port = self._update_port( [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] _ensure_no_port_binding_failure(port) [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] raise exception.PortBindingFailed(port_id=port['id']) [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] nova.exception.PortBindingFailed: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. [ 547.689740] env[62066]: ERROR nova.compute.manager [instance: 6e2496af-4d60-4315-9816-d0456458bc13] [ 547.690041] env[62066]: DEBUG nova.compute.utils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 547.692130] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Build of instance 6e2496af-4d60-4315-9816-d0456458bc13 was re-scheduled: Binding failed for port 98d36abe-ea03-453f-852d-4b4723e0ab51, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.692623] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.692883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.693073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquired lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.693459] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.694552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.711s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.696634] env[62066]: INFO nova.compute.claims [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.932654] env[62066]: DEBUG nova.network.neutron [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.168538] env[62066]: ERROR nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 548.168538] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.168538] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.168538] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.168538] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.168538] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.168538] env[62066]: ERROR nova.compute.manager raise self.value [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.168538] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.168538] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.168538] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.169638] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.169638] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.169638] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 548.169638] env[62066]: ERROR nova.compute.manager [ 548.169638] env[62066]: Traceback (most recent call last): [ 548.169638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.169638] env[62066]: listener.cb(fileno) [ 548.169638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.169638] env[62066]: result = function(*args, **kwargs) [ 548.169638] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.169638] env[62066]: return func(*args, **kwargs) [ 548.169638] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.169638] env[62066]: raise e [ 548.169638] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.169638] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 548.169638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.169638] env[62066]: created_port_ids = self._update_ports_for_instance( [ 548.169638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.169638] env[62066]: with excutils.save_and_reraise_exception(): [ 548.169638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.169638] env[62066]: self.force_reraise() [ 548.169638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.169638] env[62066]: raise self.value [ 548.169638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.169638] env[62066]: updated_port = self._update_port( [ 548.169638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.169638] env[62066]: _ensure_no_port_binding_failure(port) [ 548.169638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.169638] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.170390] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 548.170390] env[62066]: Removing descriptor: 19 [ 548.170390] env[62066]: ERROR nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Traceback (most recent call last): [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] yield resources [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.driver.spawn(context, instance, image_meta, [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.170390] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] vm_ref = self.build_virtual_machine(instance, [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] for vif in network_info: [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self._sync_wrapper(fn, *args, **kwargs) [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.wait() [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self[:] = self._gt.wait() [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self._exit_event.wait() [ 548.170831] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] result = hub.switch() [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self.greenlet.switch() [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] result = function(*args, **kwargs) [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return func(*args, **kwargs) [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise e [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] nwinfo = self.network_api.allocate_for_instance( [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.171172] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] created_port_ids = self._update_ports_for_instance( [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] with excutils.save_and_reraise_exception(): [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.force_reraise() [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise self.value [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] updated_port = self._update_port( [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] _ensure_no_port_binding_failure(port) [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.171492] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise exception.PortBindingFailed(port_id=port['id']) [ 548.171792] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 548.171792] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] [ 548.171792] env[62066]: INFO nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Terminating instance [ 548.172982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquiring lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.172982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquired lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.172982] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.230456] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.433918] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.439088] env[62066]: INFO nova.compute.manager [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] Took 1.04 seconds to deallocate network for instance. [ 548.668173] env[62066]: DEBUG nova.compute.manager [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Received event network-changed-225dc014-1d6f-42f8-a368-c3757c9165b1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.668173] env[62066]: DEBUG nova.compute.manager [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Refreshing instance network info cache due to event network-changed-225dc014-1d6f-42f8-a368-c3757c9165b1. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.668173] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] Acquiring lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.714148] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.898276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a35076-289f-4bd1-9806-6386d8557b37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.910879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4c586b-ca64-4620-ac8e-6f72c0c4532d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.952256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Releasing lock "refresh_cache-6e2496af-4d60-4315-9816-d0456458bc13" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.953031] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.953400] env[62066]: DEBUG nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.953658] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.961909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fb4d3d-b825-486b-843f-8d2b7a79c5af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.972381] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef00b9f-97fe-49d9-84fc-5b1f62efd51a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.990648] env[62066]: DEBUG nova.compute.provider_tree [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.005871] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.197371] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.484486] env[62066]: INFO nova.scheduler.client.report [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Deleted allocations for instance ed41309e-0253-4bbf-acd5-e1030ce2aed0 [ 549.492619] env[62066]: DEBUG nova.scheduler.client.report [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.510088] env[62066]: DEBUG nova.network.neutron [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.700568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Releasing lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.701036] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 549.701237] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 549.701559] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] Acquired lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.702707] env[62066]: DEBUG nova.network.neutron [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Refreshing network info cache for port 225dc014-1d6f-42f8-a368-c3757c9165b1 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 549.704725] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17b36025-7cf8-4b49-9288-f76341b5b58b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.721927] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c3be3a-a356-4f10-862e-8503b4dca632 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.754069] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a87ad1d1-9613-4144-9390-d5e3da3a2806 could not be found. [ 549.754235] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.754422] env[62066]: INFO nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Took 0.05 seconds to destroy the instance on the hypervisor. [ 549.754693] env[62066]: DEBUG oslo.service.loopingcall [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.754920] env[62066]: DEBUG nova.compute.manager [-] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.756028] env[62066]: DEBUG nova.network.neutron [-] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.814619] env[62066]: DEBUG nova.network.neutron [-] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.831541] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "30f7baa5-0288-4ef3-9d1c-d7143868b3a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.831980] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "30f7baa5-0288-4ef3-9d1c-d7143868b3a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.976605] env[62066]: ERROR nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 549.976605] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.976605] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.976605] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.976605] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.976605] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.976605] env[62066]: ERROR nova.compute.manager raise self.value [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.976605] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.976605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.976605] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.977147] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.977147] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.977147] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 549.977147] env[62066]: ERROR nova.compute.manager [ 549.977147] env[62066]: Traceback (most recent call last): [ 549.977147] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.977147] env[62066]: listener.cb(fileno) [ 549.977147] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.977147] env[62066]: result = function(*args, **kwargs) [ 549.977147] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.977147] env[62066]: return func(*args, **kwargs) [ 549.977147] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.977147] env[62066]: raise e [ 549.977147] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.977147] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 549.977147] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.977147] env[62066]: created_port_ids = self._update_ports_for_instance( [ 549.977147] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.977147] env[62066]: with excutils.save_and_reraise_exception(): [ 549.977147] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.977147] env[62066]: self.force_reraise() [ 549.977147] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.977147] env[62066]: raise self.value [ 549.977147] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.977147] env[62066]: updated_port = self._update_port( [ 549.977147] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.977147] env[62066]: _ensure_no_port_binding_failure(port) [ 549.977147] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.977147] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.978821] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 549.978821] env[62066]: Removing descriptor: 18 [ 549.978821] env[62066]: ERROR nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Traceback (most recent call last): [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] yield resources [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.driver.spawn(context, instance, image_meta, [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.978821] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] vm_ref = self.build_virtual_machine(instance, [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] for vif in network_info: [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self._sync_wrapper(fn, *args, **kwargs) [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.wait() [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self[:] = self._gt.wait() [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self._exit_event.wait() [ 549.979436] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] result = hub.switch() [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self.greenlet.switch() [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] result = function(*args, **kwargs) [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return func(*args, **kwargs) [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise e [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] nwinfo = self.network_api.allocate_for_instance( [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.979842] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] created_port_ids = self._update_ports_for_instance( [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] with excutils.save_and_reraise_exception(): [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.force_reraise() [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise self.value [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] updated_port = self._update_port( [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] _ensure_no_port_binding_failure(port) [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.980266] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise exception.PortBindingFailed(port_id=port['id']) [ 549.980610] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 549.980610] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] [ 549.980610] env[62066]: INFO nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Terminating instance [ 549.981604] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquiring lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.981720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquired lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.981894] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.997539] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.997985] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.001516] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59248c52-b735-4f4c-9ff8-6f7540702fe1 tempest-ListServerFiltersTestJSON-939767107 tempest-ListServerFiltersTestJSON-939767107-project-member] Lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.912s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.002066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.805s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.003853] env[62066]: INFO nova.compute.claims [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.007944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.703s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.008196] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ed41309e-0253-4bbf-acd5-e1030ce2aed0] During sync_power_state the instance has a pending task (spawning). Skip. [ 550.008602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "ed41309e-0253-4bbf-acd5-e1030ce2aed0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.015969] env[62066]: INFO nova.compute.manager [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] Took 1.06 seconds to deallocate network for instance. [ 550.248367] env[62066]: DEBUG nova.network.neutron [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.319988] env[62066]: DEBUG nova.network.neutron [-] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.408812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquiring lock "de135f73-53d2-4471-a0ba-9204af726690" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.408812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "de135f73-53d2-4471-a0ba-9204af726690" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.418801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquiring lock "c8bad79f-9bf7-4ef2-b681-ca8901e878fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.419071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "c8bad79f-9bf7-4ef2-b681-ca8901e878fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.456967] env[62066]: DEBUG nova.network.neutron [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.504483] env[62066]: DEBUG nova.compute.utils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 550.510018] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.510018] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 550.510018] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.518323] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.822376] env[62066]: INFO nova.compute.manager [-] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Took 1.07 seconds to deallocate network for instance. [ 550.828669] env[62066]: DEBUG nova.compute.claims [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 550.829549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.892564] env[62066]: DEBUG nova.policy [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dce5e9b642424ef5b202eb2295a3fe1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7dba0273591846a0a9522614b16bce62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.963294] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd5e22be-66b4-4f6e-b7c4-1adb326845df req-7d0b4183-7c6e-48d7-baee-a5336b9bce83 service nova] Releasing lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.001573] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.010958] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.041350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.234873] env[62066]: INFO nova.scheduler.client.report [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Deleted allocations for instance 6e2496af-4d60-4315-9816-d0456458bc13 [ 551.459233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquiring lock "80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.459233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.478729] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13294a0a-2831-4f3e-adcf-18f38909a7ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.492755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73632484-e83a-4c0d-babb-baf75db41c36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.531623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Releasing lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.532067] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 551.532280] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.536416] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c263a8f-7f98-4ef7-b11b-2a4ed617d70b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.539227] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3b3cdf-a3e4-4f1d-a6fb-8ce1e31744d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.557523] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc604e7b-c4de-4be9-b9dc-3994f6118152 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.572733] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2966e4-2411-482f-ab85-19ae97dcae79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.581640] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e could not be found. [ 551.584034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.584034] env[62066]: INFO nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 551.584034] env[62066]: DEBUG oslo.service.loopingcall [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.592029] env[62066]: DEBUG nova.compute.manager [-] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.592029] env[62066]: DEBUG nova.network.neutron [-] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.596362] env[62066]: DEBUG nova.compute.provider_tree [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.606731] env[62066]: DEBUG nova.compute.manager [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Received event network-vif-deleted-225dc014-1d6f-42f8-a368-c3757c9165b1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.607034] env[62066]: DEBUG nova.compute.manager [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Received event network-changed-1a63f849-d8b2-4293-93c6-9559584e6ed3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.607238] env[62066]: DEBUG nova.compute.manager [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Refreshing instance network info cache due to event network-changed-1a63f849-d8b2-4293-93c6-9559584e6ed3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 551.607654] env[62066]: DEBUG oslo_concurrency.lockutils [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] Acquiring lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.607654] env[62066]: DEBUG oslo_concurrency.lockutils [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] Acquired lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.607986] env[62066]: DEBUG nova.network.neutron [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Refreshing network info cache for port 1a63f849-d8b2-4293-93c6-9559584e6ed3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 551.649366] env[62066]: DEBUG nova.network.neutron [-] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.750977] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a68d288b-193d-4321-b8b4-997957481fb9 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "6e2496af-4d60-4315-9816-d0456458bc13" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.519s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.754730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "6e2496af-4d60-4315-9816-d0456458bc13" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.448s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.754730] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6e2496af-4d60-4315-9816-d0456458bc13] During sync_power_state the instance has a pending task (spawning). Skip. [ 551.754730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "6e2496af-4d60-4315-9816-d0456458bc13" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.049565] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.081399] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Successfully created port: 68344ea5-13d1-43be-ab81-bc8ef27fa804 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.089869] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.090061] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.090256] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.090457] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.090602] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.090776] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.093092] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.093589] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.093589] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.093761] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.093957] env[62066]: DEBUG nova.virt.hardware [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.094935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6948c1-7799-473c-87d1-25dad4826ad1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.098895] env[62066]: DEBUG nova.scheduler.client.report [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.110832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e70ef7-601b-467f-ba7c-496138796e93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.151319] env[62066]: DEBUG nova.network.neutron [-] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.162983] env[62066]: DEBUG nova.network.neutron [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.254608] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.516318] env[62066]: DEBUG nova.network.neutron [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.607433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.607971] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 552.612334] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.007s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.657581] env[62066]: INFO nova.compute.manager [-] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Took 1.06 seconds to deallocate network for instance. [ 552.660079] env[62066]: DEBUG nova.compute.claims [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.660079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.793748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.021904] env[62066]: DEBUG oslo_concurrency.lockutils [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] Releasing lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.022960] env[62066]: DEBUG nova.compute.manager [req-a9764445-bedb-4e80-bd9a-7b019dd24525 req-3a50b426-f70c-4930-8ec5-36f94878c6b8 service nova] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Received event network-vif-deleted-1a63f849-d8b2-4293-93c6-9559584e6ed3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.125752] env[62066]: DEBUG nova.compute.utils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 553.129055] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 553.635307] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 553.692124] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a87ad1d1-9613-4144-9390-d5e3da3a2806 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 553.694839] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 553.694839] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance fc176994-95b2-44a3-becf-6c224476851a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 553.694839] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance fa73770f-784a-44c3-ba65-e7176180760b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.196541] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.649154] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 554.687450] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.687791] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.688161] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.689609] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.689609] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.689609] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.689609] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.689609] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.690481] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.690481] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.690481] env[62066]: DEBUG nova.virt.hardware [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.691918] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a5fd88-278c-4166-9868-bc7a805a2f53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.699979] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance bdf32142-16bd-429a-ac8b-e0aa0c47b5d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.705745] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350c138e-5809-4a96-84ec-44290d368245 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.720189] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 554.729571] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 554.730770] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74687f6e-06c8-402e-9a51-55058c23a357 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.742896] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Created folder: OpenStack in parent group-v4. [ 554.743276] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating folder: Project (79a79ea7333142509a393e5ddd75b0e7). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 554.743412] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36415ed2-fae7-475d-9532-e29d1d907b73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.752420] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Created folder: Project (79a79ea7333142509a393e5ddd75b0e7) in parent group-v251573. [ 554.752603] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating folder: Instances. Parent ref: group-v251574. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 554.752827] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de8d15cb-d8ed-4691-97c4-c8fa03931a4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.761482] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Created folder: Instances in parent group-v251574. [ 554.762110] env[62066]: DEBUG oslo.service.loopingcall [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.762110] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 554.762313] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67cac7ad-3f10-43c8-9cee-51503fb7f521 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.783142] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 554.783142] env[62066]: value = "task-1155648" [ 554.783142] env[62066]: _type = "Task" [ 554.783142] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.794329] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155648, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.803833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.803833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.209463] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 92ca97dd-cd7c-4612-85d9-a05f5a785d6f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.300475] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155648, 'name': CreateVM_Task, 'duration_secs': 0.380053} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.301585] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 555.302411] env[62066]: DEBUG oslo_vmware.service [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bcf57e-57eb-45b4-8faf-fbddfe5fe378 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.310026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.310136] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.311574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 555.312093] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb0a97b2-aef4-4835-b90d-0e0c31b32812 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.316821] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 555.316821] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526195aa-f559-696d-6b5a-8194995fe330" [ 555.316821] env[62066]: _type = "Task" [ 555.316821] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.331394] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526195aa-f559-696d-6b5a-8194995fe330, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.713611] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.831560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.831974] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 555.832324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.832421] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.832959] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.833271] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85e1c1cd-8936-4a29-9fb1-3612ba59d967 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.855430] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.855628] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 555.856572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df994d5f-cf70-4a53-8613-403b6617e956 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.867555] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13488806-b66e-4116-bf1c-4049a9d4b4fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.874011] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 555.874011] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5295d2fc-2690-10ad-3c3d-3f393cf375df" [ 555.874011] env[62066]: _type = "Task" [ 555.874011] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.883803] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5295d2fc-2690-10ad-3c3d-3f393cf375df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.003991] env[62066]: ERROR nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 556.003991] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.003991] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.003991] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.003991] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.003991] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.003991] env[62066]: ERROR nova.compute.manager raise self.value [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.003991] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.003991] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.003991] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.004736] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.004736] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.004736] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 556.004736] env[62066]: ERROR nova.compute.manager [ 556.004736] env[62066]: Traceback (most recent call last): [ 556.004736] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.004736] env[62066]: listener.cb(fileno) [ 556.004736] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.004736] env[62066]: result = function(*args, **kwargs) [ 556.004736] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.004736] env[62066]: return func(*args, **kwargs) [ 556.004736] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.004736] env[62066]: raise e [ 556.004736] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.004736] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 556.004736] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.004736] env[62066]: created_port_ids = self._update_ports_for_instance( [ 556.004736] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.004736] env[62066]: with excutils.save_and_reraise_exception(): [ 556.004736] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.004736] env[62066]: self.force_reraise() [ 556.004736] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.004736] env[62066]: raise self.value [ 556.004736] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.004736] env[62066]: updated_port = self._update_port( [ 556.004736] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.004736] env[62066]: _ensure_no_port_binding_failure(port) [ 556.004736] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.004736] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.005612] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 556.005612] env[62066]: Removing descriptor: 19 [ 556.005612] env[62066]: ERROR nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] Traceback (most recent call last): [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] yield resources [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.driver.spawn(context, instance, image_meta, [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.005612] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] vm_ref = self.build_virtual_machine(instance, [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] for vif in network_info: [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self._sync_wrapper(fn, *args, **kwargs) [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.wait() [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self[:] = self._gt.wait() [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self._exit_event.wait() [ 556.005974] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] result = hub.switch() [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self.greenlet.switch() [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] result = function(*args, **kwargs) [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return func(*args, **kwargs) [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise e [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] nwinfo = self.network_api.allocate_for_instance( [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.006431] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] created_port_ids = self._update_ports_for_instance( [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] with excutils.save_and_reraise_exception(): [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.force_reraise() [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise self.value [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] updated_port = self._update_port( [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] _ensure_no_port_binding_failure(port) [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.006838] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise exception.PortBindingFailed(port_id=port['id']) [ 556.007206] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 556.007206] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] [ 556.007206] env[62066]: INFO nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Terminating instance [ 556.008571] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.008772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.008969] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.218643] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance b3d0de63-fe22-4bc7-9602-509054dca586 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.385611] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 556.385951] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating directory with path [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 556.386392] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d124bb78-c827-4cc9-86fb-8c833ecebc2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.422144] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Created directory with path [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 556.422144] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Fetch image to [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 556.422144] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Downloading image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk on the data store datastore2 {{(pid=62066) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 556.422144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20249070-e59b-4694-b0b3-ba442cc84815 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.432638] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fa66dc-2b94-4c83-8c0d-ff66eba3acbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.443804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765b1b23-6986-4e8d-a244-656ecadd3b75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.485148] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99056700-f254-4c89-bbe8-c1d152bc3c31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.493063] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4d76795a-d4cb-4c1d-96fa-e9327e9ad053 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.573172] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.587377] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Downloading image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to the data store datastore2 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 556.700675] env[62066]: DEBUG oslo_vmware.rw_handles [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 556.783761] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance cf3dc6c9-d7bf-4319-9420-e3634b9fca11 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.792609] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.977479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "b5610b23-828b-4ece-b18a-4bfe27a60bbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.977731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "b5610b23-828b-4ece-b18a-4bfe27a60bbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.186953] env[62066]: DEBUG nova.compute.manager [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Received event network-changed-68344ea5-13d1-43be-ab81-bc8ef27fa804 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.187129] env[62066]: DEBUG nova.compute.manager [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Refreshing instance network info cache due to event network-changed-68344ea5-13d1-43be-ab81-bc8ef27fa804. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.187295] env[62066]: DEBUG oslo_concurrency.lockutils [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] Acquiring lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.274723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "5fa586bb-bf1c-4126-a2a7-70e488307794" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.274723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "5fa586bb-bf1c-4126-a2a7-70e488307794" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.293314] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 30f7baa5-0288-4ef3-9d1c-d7143868b3a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.301779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.302809] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.302809] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.303087] env[62066]: DEBUG oslo_concurrency.lockutils [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] Acquired lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.303299] env[62066]: DEBUG nova.network.neutron [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Refreshing network info cache for port 68344ea5-13d1-43be-ab81-bc8ef27fa804 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 557.306836] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02d9fbee-1cd5-4e72-a920-b08b3bbf5d4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.317351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7763858c-2ceb-43de-ba00-12aa55a91b85 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.345454] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fc176994-95b2-44a3-becf-6c224476851a could not be found. [ 557.345690] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.345865] env[62066]: INFO nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 557.346170] env[62066]: DEBUG oslo.service.loopingcall [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.346448] env[62066]: DEBUG nova.compute.manager [-] [instance: fc176994-95b2-44a3-becf-6c224476851a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.346672] env[62066]: DEBUG nova.network.neutron [-] [instance: fc176994-95b2-44a3-becf-6c224476851a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.385668] env[62066]: DEBUG nova.network.neutron [-] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.479222] env[62066]: DEBUG oslo_vmware.rw_handles [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 557.479469] env[62066]: DEBUG oslo_vmware.rw_handles [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 557.538072] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Downloaded image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk on the data store datastore2 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 557.539978] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 557.540247] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Copying Virtual Disk [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk to [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 557.540870] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-716198ba-037f-4795-bd37-9ec4be839275 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.549111] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 557.549111] env[62066]: value = "task-1155649" [ 557.549111] env[62066]: _type = "Task" [ 557.549111] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.564625] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.787634] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.787634] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.795959] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance c8bad79f-9bf7-4ef2-b681-ca8901e878fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.831634] env[62066]: DEBUG nova.network.neutron [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.887850] env[62066]: DEBUG nova.network.neutron [-] [instance: fc176994-95b2-44a3-becf-6c224476851a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.923415] env[62066]: DEBUG nova.network.neutron [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.062253] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155649, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.076602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.076602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.301084] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance de135f73-53d2-4471-a0ba-9204af726690 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.391952] env[62066]: INFO nova.compute.manager [-] [instance: fc176994-95b2-44a3-becf-6c224476851a] Took 1.04 seconds to deallocate network for instance. [ 558.397230] env[62066]: DEBUG nova.compute.claims [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.397521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.428896] env[62066]: DEBUG oslo_concurrency.lockutils [req-62243641-646d-4791-88f2-4ddb6d1dd205 req-9c59b5b2-e1a3-44d2-b164-71e72c191105 service nova] Releasing lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.565191] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155649, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.804749] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.805152] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 558.805204] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 559.062913] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155649, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.045299} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.063395] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Copied Virtual Disk [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk to [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 559.063587] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleting the datastore file [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.063823] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2eafeeeb-677b-433e-928a-5c2391ac03e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.071711] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 559.071711] env[62066]: value = "task-1155650" [ 559.071711] env[62066]: _type = "Task" [ 559.071711] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.079743] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155650, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.122101] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1a5f24-7f84-49d8-b5c3-051d837dca92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.129565] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd724f7-9f91-48cb-ac3f-efb69ddc66cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.166344] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243c7d77-0768-4213-a3de-95e893199a20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.173888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a659fb00-e6a3-4796-b335-5716f322617e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.188015] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.586622] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155650, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026595} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.586952] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 559.587744] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Moving file from [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0. {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 559.588448] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a8cbf6f6-830f-4ce8-91d0-0924dee638dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.597020] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 559.597020] env[62066]: value = "task-1155651" [ 559.597020] env[62066]: _type = "Task" [ 559.597020] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.605532] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155651, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.691992] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.105180] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155651, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.027659} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.105455] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] File moved {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 560.105647] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Cleaning up location [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 560.105876] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleting the datastore file [datastore2] vmware_temp/b176e0f0-6bd9-497a-a43d-3f5f1904a688 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 560.106172] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59bffab2-2780-4dc1-86df-164f58414d31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.112953] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 560.112953] env[62066]: value = "task-1155652" [ 560.112953] env[62066]: _type = "Task" [ 560.112953] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.120868] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155652, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.205031] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 560.205031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.590s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.205031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.839s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.205031] env[62066]: INFO nova.compute.claims [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.210017] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 560.210371] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Getting list of instances from cluster (obj){ [ 560.210371] env[62066]: value = "domain-c8" [ 560.210371] env[62066]: _type = "ClusterComputeResource" [ 560.210371] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 560.211974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5c7a01-c2ae-4e93-b6dc-7aa860647144 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.222874] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Got total of 1 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 560.387275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "58528227-a9dc-4c12-b15b-fa191bc29046" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.387275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "58528227-a9dc-4c12-b15b-fa191bc29046" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.631324] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155652, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027555} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.631588] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 560.636569] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2929d895-a368-487e-86f3-c4325fb94bbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.648228] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 560.648228] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527afe54-f975-ffe4-e74c-cc8f48f8c0df" [ 560.648228] env[62066]: _type = "Task" [ 560.648228] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.666597] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527afe54-f975-ffe4-e74c-cc8f48f8c0df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.105217] env[62066]: DEBUG nova.compute.manager [req-f785b28d-912c-465b-92fe-d132d5a6d21d req-75a5f6c4-9f47-4ee6-b23b-bea8c777cbfb service nova] [instance: fc176994-95b2-44a3-becf-6c224476851a] Received event network-vif-deleted-68344ea5-13d1-43be-ab81-bc8ef27fa804 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.165060] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527afe54-f975-ffe4-e74c-cc8f48f8c0df, 'name': SearchDatastore_Task, 'duration_secs': 0.034096} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.165360] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.165668] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] fa73770f-784a-44c3-ba65-e7176180760b/fa73770f-784a-44c3-ba65-e7176180760b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 561.165981] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-380f2b37-6816-4fe1-9958-c815cd878250 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.175462] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 561.175462] env[62066]: value = "task-1155653" [ 561.175462] env[62066]: _type = "Task" [ 561.175462] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.184414] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155653, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.297070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "a406e14c-b281-4d1f-97d0-3219c6ffaa8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.298024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "a406e14c-b281-4d1f-97d0-3219c6ffaa8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.613936] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c8ed50-ad20-4364-aa43-ba34bd6af4da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.624583] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361af2be-9225-4ecb-8217-ade93c1c6146 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.670364] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a137f86a-49bc-481b-b83e-b2d4075796e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.683583] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bddcae6-b7a9-4ecf-8b5a-2cd90ba18540 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.694591] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155653, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497905} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.702592] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] fa73770f-784a-44c3-ba65-e7176180760b/fa73770f-784a-44c3-ba65-e7176180760b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 561.702820] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 561.703343] env[62066]: DEBUG nova.compute.provider_tree [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.704570] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-130543d0-7b2b-41ed-a5f0-7d7d5ea7deec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.715533] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 561.715533] env[62066]: value = "task-1155654" [ 561.715533] env[62066]: _type = "Task" [ 561.715533] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.730166] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155654, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.212202] env[62066]: DEBUG nova.scheduler.client.report [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.229438] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155654, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065114} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.229725] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 562.231335] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a22aa65-b895-41ec-b204-e8e4d144bdc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.257198] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] fa73770f-784a-44c3-ba65-e7176180760b/fa73770f-784a-44c3-ba65-e7176180760b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 562.257198] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea5b3e5b-b7a7-4b7d-b99f-c24fbc3c5013 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.277678] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 562.277678] env[62066]: value = "task-1155655" [ 562.277678] env[62066]: _type = "Task" [ 562.277678] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.286697] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155655, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.322347] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquiring lock "d347bf40-941c-428f-85c3-18067a21d6c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.322588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "d347bf40-941c-428f-85c3-18067a21d6c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.725570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.726028] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 562.731088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.972s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.732314] env[62066]: INFO nova.compute.claims [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.791949] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155655, 'name': ReconfigVM_Task, 'duration_secs': 0.279531} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.792388] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Reconfigured VM instance instance-0000000b to attach disk [datastore2] fa73770f-784a-44c3-ba65-e7176180760b/fa73770f-784a-44c3-ba65-e7176180760b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 562.792979] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87b2e1fe-6dbc-46de-820f-80609c5c7ccb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.799983] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 562.799983] env[62066]: value = "task-1155656" [ 562.799983] env[62066]: _type = "Task" [ 562.799983] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.812468] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155656, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.240423] env[62066]: DEBUG nova.compute.utils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 563.241915] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.242127] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 563.291016] env[62066]: DEBUG nova.policy [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79603f93c8884d17bdda55c8ec1a3647', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e22fc89b67b845f798c258ddbcf53cd8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 563.316361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.316638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.316854] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155656, 'name': Rename_Task, 'duration_secs': 0.12628} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.317105] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 563.317352] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36eb030c-c1b0-4eb4-8979-2bbbb124403f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.326756] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 563.326756] env[62066]: value = "task-1155657" [ 563.326756] env[62066]: _type = "Task" [ 563.326756] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.342630] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.626479] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Successfully created port: 2c3f593c-19f7-4f71-aeb2-ade6950ddd19 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.748899] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 563.839716] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155657, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.134776] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cc404f-3830-48fb-93d8-6ae948b886be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.143472] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8598f4d-6cf8-4fd4-8323-c49ba0f96c7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.186513] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c22fe6f-f0d6-49c6-8a99-feccc86fe4c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.193787] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c1d2e7-ccd9-4f62-afec-895ed1213986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.211179] env[62066]: DEBUG nova.compute.provider_tree [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.340562] env[62066]: DEBUG oslo_vmware.api [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155657, 'name': PowerOnVM_Task, 'duration_secs': 0.616747} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.340934] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 564.341115] env[62066]: INFO nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Took 9.69 seconds to spawn the instance on the hypervisor. [ 564.341402] env[62066]: DEBUG nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 564.342073] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c895ce0-c095-4755-adde-71fde9cdb4a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.592225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquiring lock "e0b6deda-1870-44f3-b39a-e971d592be9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.592511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "e0b6deda-1870-44f3-b39a-e971d592be9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.700356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquiring lock "ada6c82c-c4fd-41a2-b30c-3aa505af6383" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.700356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "ada6c82c-c4fd-41a2-b30c-3aa505af6383" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.715031] env[62066]: ERROR nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 564.715031] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.715031] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.715031] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.715031] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.715031] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.715031] env[62066]: ERROR nova.compute.manager raise self.value [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.715031] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.715031] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.715031] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.715571] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.715571] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.715571] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 564.715571] env[62066]: ERROR nova.compute.manager [ 564.715571] env[62066]: Traceback (most recent call last): [ 564.715571] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.715571] env[62066]: listener.cb(fileno) [ 564.715571] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.715571] env[62066]: result = function(*args, **kwargs) [ 564.715571] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.715571] env[62066]: return func(*args, **kwargs) [ 564.715571] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.715571] env[62066]: raise e [ 564.715571] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.715571] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 564.715571] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.715571] env[62066]: created_port_ids = self._update_ports_for_instance( [ 564.715571] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.715571] env[62066]: with excutils.save_and_reraise_exception(): [ 564.715571] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.715571] env[62066]: self.force_reraise() [ 564.715571] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.715571] env[62066]: raise self.value [ 564.715571] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.715571] env[62066]: updated_port = self._update_port( [ 564.715571] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.715571] env[62066]: _ensure_no_port_binding_failure(port) [ 564.715571] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.715571] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.716469] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 564.716469] env[62066]: Removing descriptor: 19 [ 564.716469] env[62066]: DEBUG nova.scheduler.client.report [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.765233] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 564.795334] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:52:46Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='176669148',id=31,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1989802941',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 564.795646] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 564.796209] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 564.796447] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 564.796601] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 564.796763] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 564.797250] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 564.797475] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 564.797676] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 564.798205] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 564.798680] env[62066]: DEBUG nova.virt.hardware [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 564.799558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eda27ca-e4b5-47a4-8b9b-98e2fbdee1a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.808518] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf47cd95-6565-4b24-8248-37600a412b67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.826555] env[62066]: ERROR nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Traceback (most recent call last): [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] yield resources [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.driver.spawn(context, instance, image_meta, [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] vm_ref = self.build_virtual_machine(instance, [ 564.826555] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] for vif in network_info: [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return self._sync_wrapper(fn, *args, **kwargs) [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.wait() [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self[:] = self._gt.wait() [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return self._exit_event.wait() [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 564.826996] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] current.throw(*self._exc) [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] result = function(*args, **kwargs) [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return func(*args, **kwargs) [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise e [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] nwinfo = self.network_api.allocate_for_instance( [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] created_port_ids = self._update_ports_for_instance( [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] with excutils.save_and_reraise_exception(): [ 564.827399] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.force_reraise() [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise self.value [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] updated_port = self._update_port( [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] _ensure_no_port_binding_failure(port) [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise exception.PortBindingFailed(port_id=port['id']) [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 564.827923] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] [ 564.827923] env[62066]: INFO nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Terminating instance [ 564.831112] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquiring lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.831112] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquired lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.831112] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 564.866477] env[62066]: INFO nova.compute.manager [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Took 25.70 seconds to build instance. [ 565.189782] env[62066]: DEBUG nova.compute.manager [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Received event network-changed-2c3f593c-19f7-4f71-aeb2-ade6950ddd19 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.189985] env[62066]: DEBUG nova.compute.manager [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Refreshing instance network info cache due to event network-changed-2c3f593c-19f7-4f71-aeb2-ade6950ddd19. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.191619] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] Acquiring lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.220667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.220667] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.230020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.241s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.230020] env[62066]: INFO nova.compute.claims [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.353567] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.370083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68facff3-a00a-4acf-9237-20e63465e839 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.207s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.429927] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.738990] env[62066]: DEBUG nova.compute.utils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 565.740486] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.740652] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 565.793279] env[62066]: DEBUG nova.policy [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06d9811c0388479aa07eddda926364b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9bd1a372940141cd9b085a1fc0011d9b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.873529] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.932513] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Releasing lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.933074] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.933074] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.933338] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] Acquired lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.933673] env[62066]: DEBUG nova.network.neutron [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Refreshing network info cache for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 565.934768] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b543f100-385a-4e3d-8fe4-473a090969d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.947596] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6215b55-1c93-45ac-a928-4a24c9519995 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.972994] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd could not be found. [ 565.973263] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.973412] env[62066]: INFO nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.973683] env[62066]: DEBUG oslo.service.loopingcall [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.973878] env[62066]: DEBUG nova.compute.manager [-] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.973983] env[62066]: DEBUG nova.network.neutron [-] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.995492] env[62066]: DEBUG nova.network.neutron [-] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.157867] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Successfully created port: 6894146e-8eb2-4de3-8e71-cd02b26a5c14 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 566.246428] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.400922] env[62066]: DEBUG nova.compute.manager [None req-2c51b9f0-ca4f-451f-9d1e-4d7e46974ca7 tempest-ServerDiagnosticsV248Test-1990096989 tempest-ServerDiagnosticsV248Test-1990096989-project-admin] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 566.402342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1643ae-1390-47bd-9e3e-cdad2970e90a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.407289] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.411711] env[62066]: INFO nova.compute.manager [None req-2c51b9f0-ca4f-451f-9d1e-4d7e46974ca7 tempest-ServerDiagnosticsV248Test-1990096989 tempest-ServerDiagnosticsV248Test-1990096989-project-admin] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Retrieving diagnostics [ 566.414250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8cf027-c05d-46fe-913f-de6d4ddc50d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.471883] env[62066]: DEBUG nova.network.neutron [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.500496] env[62066]: DEBUG nova.network.neutron [-] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.586166] env[62066]: DEBUG nova.network.neutron [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.690989] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefed638-d520-43fc-90e6-bc31c454a911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.699250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fc7794-352a-48cf-a26d-441ba448387e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.734038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530c682d-ca44-4e37-9678-e5e596a19c7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.741879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e5eb10-27cc-40a6-ab93-8df9c6f8c506 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.760537] env[62066]: DEBUG nova.compute.provider_tree [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.002810] env[62066]: INFO nova.compute.manager [-] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Took 1.03 seconds to deallocate network for instance. [ 567.006703] env[62066]: DEBUG nova.compute.claims [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.006703] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.089945] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67b2292-df64-4196-b4f4-c7a7704cb692 req-d1e8980b-5991-4abc-a193-8617c915417f service nova] Releasing lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.266121] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.266739] env[62066]: DEBUG nova.scheduler.client.report [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.296998] env[62066]: ERROR nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 567.296998] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.296998] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.296998] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.296998] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.296998] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.296998] env[62066]: ERROR nova.compute.manager raise self.value [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.296998] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 567.296998] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.296998] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 567.297605] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.297605] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 567.297605] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 567.297605] env[62066]: ERROR nova.compute.manager [ 567.297605] env[62066]: Traceback (most recent call last): [ 567.297605] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 567.297605] env[62066]: listener.cb(fileno) [ 567.297605] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.297605] env[62066]: result = function(*args, **kwargs) [ 567.297605] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.297605] env[62066]: return func(*args, **kwargs) [ 567.297605] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.297605] env[62066]: raise e [ 567.297605] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.297605] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 567.297605] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.297605] env[62066]: created_port_ids = self._update_ports_for_instance( [ 567.297605] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.297605] env[62066]: with excutils.save_and_reraise_exception(): [ 567.297605] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.297605] env[62066]: self.force_reraise() [ 567.297605] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.297605] env[62066]: raise self.value [ 567.297605] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.297605] env[62066]: updated_port = self._update_port( [ 567.297605] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.297605] env[62066]: _ensure_no_port_binding_failure(port) [ 567.297605] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.297605] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 567.298609] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 567.298609] env[62066]: Removing descriptor: 19 [ 567.299705] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.299960] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.300131] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.300364] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.300516] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.300719] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.300962] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.301014] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.301310] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.301480] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.301661] env[62066]: DEBUG nova.virt.hardware [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.302894] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d383f4cb-fa52-4a13-8f63-2d4c2aead676 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.311680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2022e56-2d34-4448-8742-584b4063f442 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.327176] env[62066]: ERROR nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Traceback (most recent call last): [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] yield resources [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.driver.spawn(context, instance, image_meta, [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] vm_ref = self.build_virtual_machine(instance, [ 567.327176] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] for vif in network_info: [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return self._sync_wrapper(fn, *args, **kwargs) [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.wait() [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self[:] = self._gt.wait() [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return self._exit_event.wait() [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 567.327628] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] current.throw(*self._exc) [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] result = function(*args, **kwargs) [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return func(*args, **kwargs) [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise e [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] nwinfo = self.network_api.allocate_for_instance( [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] created_port_ids = self._update_ports_for_instance( [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] with excutils.save_and_reraise_exception(): [ 567.328208] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.force_reraise() [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise self.value [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] updated_port = self._update_port( [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] _ensure_no_port_binding_failure(port) [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise exception.PortBindingFailed(port_id=port['id']) [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 567.328739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] [ 567.328739] env[62066]: INFO nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Terminating instance [ 567.329616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquiring lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.329734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquired lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.329899] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.607929] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "3f0bf344-d408-471d-87e2-d3e0ef5c40fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.608550] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "3f0bf344-d408-471d-87e2-d3e0ef5c40fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.635838] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "9609ae16-202e-439b-a7e4-c1e7a0254a74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.636204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "9609ae16-202e-439b-a7e4-c1e7a0254a74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.712019] env[62066]: DEBUG nova.compute.manager [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Received event network-vif-deleted-2c3f593c-19f7-4f71-aeb2-ade6950ddd19 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.712541] env[62066]: DEBUG nova.compute.manager [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Received event network-changed-6894146e-8eb2-4de3-8e71-cd02b26a5c14 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.712838] env[62066]: DEBUG nova.compute.manager [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Refreshing instance network info cache due to event network-changed-6894146e-8eb2-4de3-8e71-cd02b26a5c14. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 567.713410] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] Acquiring lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.771882] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.773266] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.776244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.304s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.777967] env[62066]: INFO nova.compute.claims [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.850670] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.924959] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.282986] env[62066]: DEBUG nova.compute.utils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 568.286233] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.286412] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 568.349283] env[62066]: DEBUG nova.policy [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '755b1ff1e4914384afadf2f1bbfedde5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef83268550f5441caae3e96025166fab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.429935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Releasing lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.429935] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 568.429935] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 568.430493] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] Acquired lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.430692] env[62066]: DEBUG nova.network.neutron [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Refreshing network info cache for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 568.432484] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f7e0ab7c-6017-462e-ae5a-e12f606ff28f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.445646] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c37091-67af-47f9-a621-ef3fd3cc0cb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.471618] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bdf32142-16bd-429a-ac8b-e0aa0c47b5d0 could not be found. [ 568.471872] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 568.472073] env[62066]: INFO nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 568.472339] env[62066]: DEBUG oslo.service.loopingcall [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.472613] env[62066]: DEBUG nova.compute.manager [-] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.472708] env[62066]: DEBUG nova.network.neutron [-] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.492574] env[62066]: DEBUG nova.network.neutron [-] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.664856] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Successfully created port: 52508203-efc9-4b34-b970-4debe6b97ded {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.726168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquiring lock "e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.726467] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.789965] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.961318] env[62066]: DEBUG nova.network.neutron [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.995024] env[62066]: DEBUG nova.network.neutron [-] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.244674] env[62066]: DEBUG nova.network.neutron [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.255402] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08c29ab-2752-414c-81fb-4a068f9794d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.263947] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57790bb8-29f8-40ca-8fe5-87aeeeab60ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.296107] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98d82a0-becd-413c-9503-ce3579c0b275 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.306400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013e60fb-abd9-41fc-8e82-9af9bcb2db2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.321012] env[62066]: DEBUG nova.compute.provider_tree [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.500726] env[62066]: INFO nova.compute.manager [-] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Took 1.03 seconds to deallocate network for instance. [ 569.504687] env[62066]: DEBUG nova.compute.claims [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.504877] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.747708] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff42f7a-ac18-4798-8c8e-821ef49490da req-05d8fee2-e57a-4bfe-a027-f55d8a2921c0 service nova] Releasing lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.813956] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.823660] env[62066]: DEBUG nova.scheduler.client.report [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.840829] env[62066]: ERROR nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 569.840829] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.840829] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.840829] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.840829] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.840829] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.840829] env[62066]: ERROR nova.compute.manager raise self.value [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.840829] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.840829] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.840829] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.841531] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.841531] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.841531] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 569.841531] env[62066]: ERROR nova.compute.manager [ 569.841531] env[62066]: Traceback (most recent call last): [ 569.841531] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.841531] env[62066]: listener.cb(fileno) [ 569.841531] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.841531] env[62066]: result = function(*args, **kwargs) [ 569.841531] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.841531] env[62066]: return func(*args, **kwargs) [ 569.841531] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.841531] env[62066]: raise e [ 569.841531] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.841531] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 569.841531] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.841531] env[62066]: created_port_ids = self._update_ports_for_instance( [ 569.841531] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.841531] env[62066]: with excutils.save_and_reraise_exception(): [ 569.841531] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.841531] env[62066]: self.force_reraise() [ 569.841531] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.841531] env[62066]: raise self.value [ 569.841531] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.841531] env[62066]: updated_port = self._update_port( [ 569.841531] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.841531] env[62066]: _ensure_no_port_binding_failure(port) [ 569.841531] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.841531] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.842441] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 569.842441] env[62066]: Removing descriptor: 19 [ 569.847112] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.847354] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.848134] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.848134] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.848134] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.848134] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.848414] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.848578] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.848679] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.848835] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.849017] env[62066]: DEBUG nova.virt.hardware [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.849933] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b222e9b-93a8-480e-8216-1220bc69322a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.859853] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146c59f2-3829-4b04-95c1-b8b9b99a6918 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.875490] env[62066]: ERROR nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Traceback (most recent call last): [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] yield resources [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.driver.spawn(context, instance, image_meta, [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] vm_ref = self.build_virtual_machine(instance, [ 569.875490] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] for vif in network_info: [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return self._sync_wrapper(fn, *args, **kwargs) [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.wait() [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self[:] = self._gt.wait() [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return self._exit_event.wait() [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 569.875885] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] current.throw(*self._exc) [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] result = function(*args, **kwargs) [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return func(*args, **kwargs) [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise e [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] nwinfo = self.network_api.allocate_for_instance( [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] created_port_ids = self._update_ports_for_instance( [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] with excutils.save_and_reraise_exception(): [ 569.876287] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.force_reraise() [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise self.value [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] updated_port = self._update_port( [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] _ensure_no_port_binding_failure(port) [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise exception.PortBindingFailed(port_id=port['id']) [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 569.876727] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] [ 569.876727] env[62066]: INFO nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Terminating instance [ 569.877589] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquiring lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.877751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquired lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.877916] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.900710] env[62066]: DEBUG nova.compute.manager [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Received event network-vif-deleted-6894146e-8eb2-4de3-8e71-cd02b26a5c14 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.900918] env[62066]: DEBUG nova.compute.manager [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Received event network-changed-52508203-efc9-4b34-b970-4debe6b97ded {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.901099] env[62066]: DEBUG nova.compute.manager [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Refreshing instance network info cache due to event network-changed-52508203-efc9-4b34-b970-4debe6b97ded. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.901290] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] Acquiring lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.329707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.330672] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.333805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.407558] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.537810] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.846046] env[62066]: DEBUG nova.compute.utils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 570.850295] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.850628] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 570.933110] env[62066]: DEBUG nova.policy [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48182fb845b476c9cdf15dc99f9c1da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae83fb1cd0cf41c2ae932fa5d06ea4ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.044230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Releasing lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.044230] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.044230] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.045062] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] Acquired lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.045264] env[62066]: DEBUG nova.network.neutron [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Refreshing network info cache for port 52508203-efc9-4b34-b970-4debe6b97ded {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 571.046731] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01214e98-666f-4cab-9d64-e5da3f505cc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.056539] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ede569-0291-4556-b30f-f2ccfff9b891 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.085972] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92ca97dd-cd7c-4612-85d9-a05f5a785d6f could not be found. [ 571.086257] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.086487] env[62066]: INFO nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.086794] env[62066]: DEBUG oslo.service.loopingcall [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.089566] env[62066]: DEBUG nova.compute.manager [-] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.090160] env[62066]: DEBUG nova.network.neutron [-] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.107886] env[62066]: DEBUG nova.network.neutron [-] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.301372] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Successfully created port: c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.357031] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.364131] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89702392-0239-4bd9-a3d4-f254cc48ff30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.374707] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c2ef7d-5f92-44de-9d84-ce64ee80bf1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.423749] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971c8020-6fb1-431f-bf35-60b8d014aa04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.433276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651b45ed-02b3-4ed6-8cca-b968e13b6474 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.447960] env[62066]: DEBUG nova.compute.provider_tree [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.569213] env[62066]: DEBUG nova.network.neutron [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.610658] env[62066]: DEBUG nova.network.neutron [-] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.690118] env[62066]: DEBUG nova.network.neutron [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.951179] env[62066]: DEBUG nova.scheduler.client.report [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.112885] env[62066]: INFO nova.compute.manager [-] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Took 1.02 seconds to deallocate network for instance. [ 572.115998] env[62066]: DEBUG nova.compute.claims [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.117695] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.193212] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3dcaf56-2264-4357-98be-390a41eb5e44 req-b5ba6e9e-84b3-44c7-88bd-0a8caf4c4d22 service nova] Releasing lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.248125] env[62066]: ERROR nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 572.248125] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.248125] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.248125] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.248125] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.248125] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.248125] env[62066]: ERROR nova.compute.manager raise self.value [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.248125] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.248125] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.248125] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.248741] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.248741] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.248741] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 572.248741] env[62066]: ERROR nova.compute.manager [ 572.248741] env[62066]: Traceback (most recent call last): [ 572.248741] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.248741] env[62066]: listener.cb(fileno) [ 572.248741] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.248741] env[62066]: result = function(*args, **kwargs) [ 572.248741] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.248741] env[62066]: return func(*args, **kwargs) [ 572.248741] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.248741] env[62066]: raise e [ 572.248741] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.248741] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 572.248741] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.248741] env[62066]: created_port_ids = self._update_ports_for_instance( [ 572.248741] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.248741] env[62066]: with excutils.save_and_reraise_exception(): [ 572.248741] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.248741] env[62066]: self.force_reraise() [ 572.248741] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.248741] env[62066]: raise self.value [ 572.248741] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.248741] env[62066]: updated_port = self._update_port( [ 572.248741] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.248741] env[62066]: _ensure_no_port_binding_failure(port) [ 572.248741] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.248741] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.249655] env[62066]: nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 572.249655] env[62066]: Removing descriptor: 19 [ 572.342990] env[62066]: DEBUG nova.compute.manager [req-418268ca-b1e0-42b6-ba5b-1fbd12eaa746 req-98caac90-dbe2-4635-84cf-e4807d2798e3 service nova] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Received event network-vif-deleted-52508203-efc9-4b34-b970-4debe6b97ded {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.372549] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.403252] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.403252] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.403252] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.403469] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.403469] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.403469] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.407228] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.407415] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.407588] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.407753] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.407925] env[62066]: DEBUG nova.virt.hardware [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.408840] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4537328d-8fa4-4fe9-9788-044d24567bd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.418047] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89904fd6-c8ed-4566-975e-8efb05e871dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.433213] env[62066]: ERROR nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Traceback (most recent call last): [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] yield resources [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.driver.spawn(context, instance, image_meta, [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] vm_ref = self.build_virtual_machine(instance, [ 572.433213] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] for vif in network_info: [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return self._sync_wrapper(fn, *args, **kwargs) [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.wait() [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self[:] = self._gt.wait() [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return self._exit_event.wait() [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 572.433711] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] current.throw(*self._exc) [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] result = function(*args, **kwargs) [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return func(*args, **kwargs) [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise e [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] nwinfo = self.network_api.allocate_for_instance( [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] created_port_ids = self._update_ports_for_instance( [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] with excutils.save_and_reraise_exception(): [ 572.434171] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.force_reraise() [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise self.value [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] updated_port = self._update_port( [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] _ensure_no_port_binding_failure(port) [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise exception.PortBindingFailed(port_id=port['id']) [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 572.434602] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] [ 572.434602] env[62066]: INFO nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Terminating instance [ 572.435901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.436068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquired lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.438289] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.459648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.126s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.460273] env[62066]: ERROR nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Traceback (most recent call last): [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.driver.spawn(context, instance, image_meta, [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] vm_ref = self.build_virtual_machine(instance, [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.460273] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] for vif in network_info: [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self._sync_wrapper(fn, *args, **kwargs) [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.wait() [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self[:] = self._gt.wait() [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self._exit_event.wait() [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] result = hub.switch() [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.460693] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return self.greenlet.switch() [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] result = function(*args, **kwargs) [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] return func(*args, **kwargs) [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise e [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] nwinfo = self.network_api.allocate_for_instance( [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] created_port_ids = self._update_ports_for_instance( [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] with excutils.save_and_reraise_exception(): [ 572.461124] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] self.force_reraise() [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise self.value [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] updated_port = self._update_port( [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] _ensure_no_port_binding_failure(port) [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] raise exception.PortBindingFailed(port_id=port['id']) [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] nova.exception.PortBindingFailed: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. [ 572.461539] env[62066]: ERROR nova.compute.manager [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] [ 572.461930] env[62066]: DEBUG nova.compute.utils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 572.462202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.421s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.463875] env[62066]: INFO nova.compute.claims [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.471038] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Build of instance a87ad1d1-9613-4144-9390-d5e3da3a2806 was re-scheduled: Binding failed for port 225dc014-1d6f-42f8-a368-c3757c9165b1, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.471577] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.471810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquiring lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.471955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Acquired lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.472129] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 572.956687] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.005997] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.056646] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.101068] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.559637] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Releasing lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.559846] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.560045] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.560854] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dff63fa1-14f5-45a8-89b6-4f9763fd6aa0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.573242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11845c7a-ca30-4bba-964c-a83db919c415 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.604395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Releasing lock "refresh_cache-a87ad1d1-9613-4144-9390-d5e3da3a2806" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.605333] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.605333] env[62066]: DEBUG nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.606453] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.608119] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b could not be found. [ 573.608394] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.608579] env[62066]: INFO nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 573.609645] env[62066]: DEBUG oslo.service.loopingcall [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.612026] env[62066]: DEBUG nova.compute.manager [-] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.612026] env[62066]: DEBUG nova.network.neutron [-] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 573.635431] env[62066]: DEBUG nova.network.neutron [-] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.651224] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.987393] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a691bd5e-9e99-4fe3-8f18-c2249b6578f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.997222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dc95d1-cb74-4138-bc02-1edc6edb1e23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.030210] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c292e78f-52c7-46b0-b084-79365457c574 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.039183] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103bbb05-bda3-4afa-8b26-6505b8bb3604 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.055478] env[62066]: DEBUG nova.compute.provider_tree [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.139337] env[62066]: DEBUG nova.network.neutron [-] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.154516] env[62066]: DEBUG nova.network.neutron [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.524661] env[62066]: DEBUG nova.compute.manager [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Received event network-changed-c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.524931] env[62066]: DEBUG nova.compute.manager [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Refreshing instance network info cache due to event network-changed-c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.525254] env[62066]: DEBUG oslo_concurrency.lockutils [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] Acquiring lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.525393] env[62066]: DEBUG oslo_concurrency.lockutils [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] Acquired lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.525584] env[62066]: DEBUG nova.network.neutron [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Refreshing network info cache for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.558091] env[62066]: DEBUG nova.scheduler.client.report [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.643034] env[62066]: INFO nova.compute.manager [-] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Took 1.03 seconds to deallocate network for instance. [ 574.654417] env[62066]: DEBUG nova.compute.claims [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.654852] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.660270] env[62066]: INFO nova.compute.manager [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] [instance: a87ad1d1-9613-4144-9390-d5e3da3a2806] Took 1.05 seconds to deallocate network for instance. [ 574.786202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquiring lock "d584bc1e-b5a3-4885-a6fb-efec8267d894" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.786202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "d584bc1e-b5a3-4885-a6fb-efec8267d894" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.046150] env[62066]: DEBUG nova.network.neutron [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.064331] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.064827] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.067609] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.408s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.123386] env[62066]: DEBUG nova.network.neutron [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.572469] env[62066]: DEBUG nova.compute.utils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 575.577401] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.577808] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 575.626080] env[62066]: DEBUG oslo_concurrency.lockutils [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] Releasing lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.626377] env[62066]: DEBUG nova.compute.manager [req-3ad2fbe4-80e0-4429-a3c0-363611e791e5 req-091d2f7d-520b-4a01-82e0-c465b5f9bc9c service nova] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Received event network-vif-deleted-c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.674720] env[62066]: DEBUG nova.policy [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51e6d0ee9a724dae8b0dbad5833fa08e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e79f4c20e52848c2b3ec7b70b588c3f7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.699781] env[62066]: INFO nova.scheduler.client.report [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Deleted allocations for instance a87ad1d1-9613-4144-9390-d5e3da3a2806 [ 575.928701] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8f33a439-9016-42da-bcd2-fe2a697691a5 tempest-ServersListShow296Test-8666260 tempest-ServersListShow296Test-8666260-project-member] Acquiring lock "ee5e1fd5-a946-4f00-ac06-8b3973873418" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.928946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8f33a439-9016-42da-bcd2-fe2a697691a5 tempest-ServersListShow296Test-8666260 tempest-ServersListShow296Test-8666260-project-member] Lock "ee5e1fd5-a946-4f00-ac06-8b3973873418" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.081338] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.131063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ba7dc9-cbd2-417b-b24a-db3916f32405 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.142472] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd5b886-6471-4882-89d6-7ba84429b744 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.186353] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Successfully created port: c4e9385c-f0d1-4cb2-a486-80a4186def11 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.189254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f956aa-d96d-4b20-9ce2-ca6b3ae9288a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.199608] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa686a8-e93e-4662-861d-77fcc623606e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.218227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9bcf4936-ce57-47bb-9a53-201d01de3eb3 tempest-VolumesAssistedSnapshotsTest-1810145196 tempest-VolumesAssistedSnapshotsTest-1810145196-project-member] Lock "a87ad1d1-9613-4144-9390-d5e3da3a2806" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.434s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.218227] env[62066]: DEBUG nova.compute.provider_tree [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.723964] env[62066]: DEBUG nova.scheduler.client.report [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.730155] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.100223] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.116458] env[62066]: DEBUG nova.compute.manager [None req-ee048275-f4e6-4b91-92a4-bfc7f14f5db3 tempest-ServerDiagnosticsV248Test-1990096989 tempest-ServerDiagnosticsV248Test-1990096989-project-admin] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 577.116458] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84c83a4-5aa2-4acf-bba5-c575d9602a1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.127221] env[62066]: INFO nova.compute.manager [None req-ee048275-f4e6-4b91-92a4-bfc7f14f5db3 tempest-ServerDiagnosticsV248Test-1990096989 tempest-ServerDiagnosticsV248Test-1990096989-project-admin] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Retrieving diagnostics [ 577.128563] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8831ee1-56ab-4f39-8aab-2fda75e8c221 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.183806] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.184943] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.184943] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.184943] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.184943] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.184943] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.185171] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.185237] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.185418] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.185580] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.185769] env[62066]: DEBUG nova.virt.hardware [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.187414] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83e02f0-b2e7-46ed-aaba-de61a240b46e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.197315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1437d871-2b9f-4270-bfae-04e30edd7046 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.233656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.166s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.234447] env[62066]: ERROR nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Traceback (most recent call last): [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.driver.spawn(context, instance, image_meta, [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] vm_ref = self.build_virtual_machine(instance, [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.234447] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] for vif in network_info: [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self._sync_wrapper(fn, *args, **kwargs) [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.wait() [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self[:] = self._gt.wait() [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self._exit_event.wait() [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] result = hub.switch() [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.234889] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return self.greenlet.switch() [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] result = function(*args, **kwargs) [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] return func(*args, **kwargs) [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise e [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] nwinfo = self.network_api.allocate_for_instance( [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] created_port_ids = self._update_ports_for_instance( [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] with excutils.save_and_reraise_exception(): [ 577.235341] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] self.force_reraise() [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise self.value [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] updated_port = self._update_port( [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] _ensure_no_port_binding_failure(port) [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] raise exception.PortBindingFailed(port_id=port['id']) [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] nova.exception.PortBindingFailed: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. [ 577.235808] env[62066]: ERROR nova.compute.manager [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] [ 577.236191] env[62066]: DEBUG nova.compute.utils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 577.238814] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Build of instance 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e was re-scheduled: Binding failed for port 1a63f849-d8b2-4293-93c6-9559584e6ed3, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.239307] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.239622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquiring lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.239713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Acquired lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.239827] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.243495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.447s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.244920] env[62066]: INFO nova.compute.claims [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.268149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.296867] env[62066]: DEBUG nova.compute.manager [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Received event network-changed-c4e9385c-f0d1-4cb2-a486-80a4186def11 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.297081] env[62066]: DEBUG nova.compute.manager [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Refreshing instance network info cache due to event network-changed-c4e9385c-f0d1-4cb2-a486-80a4186def11. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 577.297290] env[62066]: DEBUG oslo_concurrency.lockutils [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] Acquiring lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.297487] env[62066]: DEBUG oslo_concurrency.lockutils [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] Acquired lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.297702] env[62066]: DEBUG nova.network.neutron [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Refreshing network info cache for port c4e9385c-f0d1-4cb2-a486-80a4186def11 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.383478] env[62066]: ERROR nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 577.383478] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.383478] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.383478] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.383478] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.383478] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.383478] env[62066]: ERROR nova.compute.manager raise self.value [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.383478] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.383478] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.383478] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.383984] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.383984] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.383984] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 577.383984] env[62066]: ERROR nova.compute.manager [ 577.383984] env[62066]: Traceback (most recent call last): [ 577.383984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.383984] env[62066]: listener.cb(fileno) [ 577.383984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.383984] env[62066]: result = function(*args, **kwargs) [ 577.383984] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.383984] env[62066]: return func(*args, **kwargs) [ 577.383984] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.383984] env[62066]: raise e [ 577.383984] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.383984] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 577.383984] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.383984] env[62066]: created_port_ids = self._update_ports_for_instance( [ 577.383984] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.383984] env[62066]: with excutils.save_and_reraise_exception(): [ 577.383984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.383984] env[62066]: self.force_reraise() [ 577.383984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.383984] env[62066]: raise self.value [ 577.383984] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.383984] env[62066]: updated_port = self._update_port( [ 577.383984] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.383984] env[62066]: _ensure_no_port_binding_failure(port) [ 577.383984] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.383984] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.384845] env[62066]: nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 577.384845] env[62066]: Removing descriptor: 18 [ 577.384845] env[62066]: ERROR nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Traceback (most recent call last): [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] yield resources [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.driver.spawn(context, instance, image_meta, [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.384845] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] vm_ref = self.build_virtual_machine(instance, [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] for vif in network_info: [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self._sync_wrapper(fn, *args, **kwargs) [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.wait() [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self[:] = self._gt.wait() [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self._exit_event.wait() [ 577.385222] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] result = hub.switch() [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self.greenlet.switch() [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] result = function(*args, **kwargs) [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return func(*args, **kwargs) [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise e [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] nwinfo = self.network_api.allocate_for_instance( [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.385579] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] created_port_ids = self._update_ports_for_instance( [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] with excutils.save_and_reraise_exception(): [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.force_reraise() [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise self.value [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] updated_port = self._update_port( [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] _ensure_no_port_binding_failure(port) [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.386027] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise exception.PortBindingFailed(port_id=port['id']) [ 577.386365] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 577.386365] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] [ 577.386365] env[62066]: INFO nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Terminating instance [ 577.388271] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquiring lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.771590] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.825338] env[62066]: DEBUG nova.network.neutron [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.884306] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.889475] env[62066]: DEBUG nova.network.neutron [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.142948] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "d6837004-9b24-45f7-9ed7-48c4503859cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.143252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.389574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Releasing lock "refresh_cache-7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.389933] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.390034] env[62066]: DEBUG nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.390180] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.392176] env[62066]: DEBUG oslo_concurrency.lockutils [req-be58b928-1be2-48ee-88d0-5d9aa933d805 req-3a257bbe-a783-48aa-b530-8d2db1b7c675 service nova] Releasing lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.396029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquired lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.396029] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 578.410985] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.452451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "fa73770f-784a-44c3-ba65-e7176180760b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.452709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.452919] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "fa73770f-784a-44c3-ba65-e7176180760b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.453113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.453282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.455994] env[62066]: INFO nova.compute.manager [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Terminating instance [ 578.457759] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "refresh_cache-fa73770f-784a-44c3-ba65-e7176180760b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.457951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquired lock "refresh_cache-fa73770f-784a-44c3-ba65-e7176180760b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.458116] env[62066]: DEBUG nova.network.neutron [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 578.656738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3689df5b-bad7-45e0-a126-55151c1c6574 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.664947] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5a2d88-73e8-4c3a-94a7-86d6fa778034 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.695677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f706fcf2-ffc9-49e1-a927-3d64e8fab1c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.703747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af95eb26-9a3b-44f9-881a-88f11bd92669 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.717475] env[62066]: DEBUG nova.compute.provider_tree [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.917851] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.919963] env[62066]: DEBUG nova.network.neutron [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.993252] env[62066]: DEBUG nova.network.neutron [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.030345] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.090169] env[62066]: DEBUG nova.network.neutron [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.223884] env[62066]: DEBUG nova.scheduler.client.report [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.335817] env[62066]: DEBUG nova.compute.manager [req-b767b778-0f9e-4faa-8f47-d78b5eb855a2 req-25a46266-bd96-4489-a0f4-8f3d34970058 service nova] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Received event network-vif-deleted-c4e9385c-f0d1-4cb2-a486-80a4186def11 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.423231] env[62066]: INFO nova.compute.manager [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] [instance: 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e] Took 1.03 seconds to deallocate network for instance. [ 579.536019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Releasing lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.536019] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.536019] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.536019] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8de79b63-7e30-4eee-9f10-3da9fba0ff72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.544332] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e179f8-e58e-4892-8070-18aee3c1cf5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.569173] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3d0de63-fe22-4bc7-9602-509054dca586 could not be found. [ 579.569559] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.569874] env[62066]: INFO nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Took 0.04 seconds to destroy the instance on the hypervisor. [ 579.570236] env[62066]: DEBUG oslo.service.loopingcall [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.570557] env[62066]: DEBUG nova.compute.manager [-] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.570766] env[62066]: DEBUG nova.network.neutron [-] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 579.587069] env[62066]: DEBUG nova.network.neutron [-] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.594035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Releasing lock "refresh_cache-fa73770f-784a-44c3-ba65-e7176180760b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.594035] env[62066]: DEBUG nova.compute.manager [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.594035] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.594035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2719d4e2-40bd-485d-b3b4-a439f1538b37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.603249] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 579.603474] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75507d7f-82ee-4ecf-8e01-3040965824f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.609877] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 579.609877] env[62066]: value = "task-1155668" [ 579.609877] env[62066]: _type = "Task" [ 579.609877] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.619385] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.730042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.730408] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.733071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.336s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.090347] env[62066]: DEBUG nova.network.neutron [-] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.123483] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155668, 'name': PowerOffVM_Task, 'duration_secs': 0.117479} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.124078] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 580.124262] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 580.124509] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb9ae36b-ddce-42f8-88c2-0a9656aac52b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.154133] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 580.154364] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 580.154556] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleting the datastore file [datastore2] fa73770f-784a-44c3-ba65-e7176180760b {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 580.154825] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d56401a-e49d-4bd3-9f11-e62f1c8e889a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.170288] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for the task: (returnval){ [ 580.170288] env[62066]: value = "task-1155670" [ 580.170288] env[62066]: _type = "Task" [ 580.170288] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.181509] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.242338] env[62066]: DEBUG nova.compute.utils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 580.247134] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.247134] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 580.337761] env[62066]: DEBUG nova.policy [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c575e8611164519b3fa7774ffdfc892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42be1b4424d540849df5f695a386f49b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.456553] env[62066]: INFO nova.scheduler.client.report [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Deleted allocations for instance 7c8ca2e2-bbc4-4822-8039-dc7b60c9596e [ 580.593602] env[62066]: INFO nova.compute.manager [-] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Took 1.02 seconds to deallocate network for instance. [ 580.596858] env[62066]: DEBUG nova.compute.claims [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.597133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.683520] env[62066]: DEBUG oslo_vmware.api [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Task: {'id': task-1155670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104416} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.683809] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 580.684008] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 580.684222] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.684406] env[62066]: INFO nova.compute.manager [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 580.684644] env[62066]: DEBUG oslo.service.loopingcall [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.687476] env[62066]: DEBUG nova.compute.manager [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.687542] env[62066]: DEBUG nova.network.neutron [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.713354] env[62066]: DEBUG nova.network.neutron [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.752659] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.823594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d0b29a-6609-4242-ba2a-e8a356cab1ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.832413] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71f5e81-8c3c-41e1-882b-d162912f5789 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.863849] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Successfully created port: 803ac5bc-080a-4690-b757-02374a841ea3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.866273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b12b3b-361f-4c8f-9003-72a170df5085 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.874773] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c2f24a-340e-4b40-9fea-1c6d2a664018 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.889571] env[62066]: DEBUG nova.compute.provider_tree [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.973389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e4f0c19-163f-4d2f-bca3-a7afd2625046 tempest-ServerTagsTestJSON-1101398432 tempest-ServerTagsTestJSON-1101398432-project-member] Lock "7c8ca2e2-bbc4-4822-8039-dc7b60c9596e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.219422] env[62066]: DEBUG nova.network.neutron [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.393022] env[62066]: DEBUG nova.scheduler.client.report [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.476490] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.721912] env[62066]: INFO nova.compute.manager [-] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Took 1.03 seconds to deallocate network for instance. [ 581.766567] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.802442] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.803388] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.803388] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.803388] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.803388] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.803388] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.803671] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.803671] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.803779] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.803932] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.804167] env[62066]: DEBUG nova.virt.hardware [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.805046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e069fa1-6ab6-492e-90f8-39f5aee0c150 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.814164] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e612448-2f8a-4be3-acb0-4ebbb2fcf096 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.897461] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.164s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.898243] env[62066]: ERROR nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] Traceback (most recent call last): [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.driver.spawn(context, instance, image_meta, [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] vm_ref = self.build_virtual_machine(instance, [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.898243] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] for vif in network_info: [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self._sync_wrapper(fn, *args, **kwargs) [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.wait() [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self[:] = self._gt.wait() [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self._exit_event.wait() [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] result = hub.switch() [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.898657] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return self.greenlet.switch() [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] result = function(*args, **kwargs) [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] return func(*args, **kwargs) [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise e [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] nwinfo = self.network_api.allocate_for_instance( [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] created_port_ids = self._update_ports_for_instance( [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] with excutils.save_and_reraise_exception(): [ 581.899104] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] self.force_reraise() [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise self.value [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] updated_port = self._update_port( [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] _ensure_no_port_binding_failure(port) [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] raise exception.PortBindingFailed(port_id=port['id']) [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] nova.exception.PortBindingFailed: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. [ 581.899595] env[62066]: ERROR nova.compute.manager [instance: fc176994-95b2-44a3-becf-6c224476851a] [ 581.899956] env[62066]: DEBUG nova.compute.utils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 581.900353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.493s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.902394] env[62066]: INFO nova.compute.claims [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.910393] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Build of instance fc176994-95b2-44a3-becf-6c224476851a was re-scheduled: Binding failed for port 68344ea5-13d1-43be-ab81-bc8ef27fa804, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 581.910393] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 581.910582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.914681] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.914681] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.011030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.229040] env[62066]: DEBUG nova.compute.manager [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Received event network-changed-803ac5bc-080a-4690-b757-02374a841ea3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.229307] env[62066]: DEBUG nova.compute.manager [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Refreshing instance network info cache due to event network-changed-803ac5bc-080a-4690-b757-02374a841ea3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 582.229828] env[62066]: DEBUG oslo_concurrency.lockutils [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] Acquiring lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.230011] env[62066]: DEBUG oslo_concurrency.lockutils [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] Acquired lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.230770] env[62066]: DEBUG nova.network.neutron [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Refreshing network info cache for port 803ac5bc-080a-4690-b757-02374a841ea3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 582.233340] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.367826] env[62066]: ERROR nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 582.367826] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.367826] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.367826] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.367826] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.367826] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.367826] env[62066]: ERROR nova.compute.manager raise self.value [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.367826] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.367826] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.367826] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.368460] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.368460] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.368460] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 582.368460] env[62066]: ERROR nova.compute.manager [ 582.368460] env[62066]: Traceback (most recent call last): [ 582.368460] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.368460] env[62066]: listener.cb(fileno) [ 582.368460] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.368460] env[62066]: result = function(*args, **kwargs) [ 582.368460] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.368460] env[62066]: return func(*args, **kwargs) [ 582.368460] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.368460] env[62066]: raise e [ 582.368460] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.368460] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 582.368460] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.368460] env[62066]: created_port_ids = self._update_ports_for_instance( [ 582.368460] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.368460] env[62066]: with excutils.save_and_reraise_exception(): [ 582.368460] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.368460] env[62066]: self.force_reraise() [ 582.368460] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.368460] env[62066]: raise self.value [ 582.368460] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.368460] env[62066]: updated_port = self._update_port( [ 582.368460] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.368460] env[62066]: _ensure_no_port_binding_failure(port) [ 582.368460] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.368460] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.369392] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 582.369392] env[62066]: Removing descriptor: 18 [ 582.369392] env[62066]: ERROR nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Traceback (most recent call last): [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] yield resources [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.driver.spawn(context, instance, image_meta, [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.369392] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] vm_ref = self.build_virtual_machine(instance, [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] for vif in network_info: [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self._sync_wrapper(fn, *args, **kwargs) [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.wait() [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self[:] = self._gt.wait() [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self._exit_event.wait() [ 582.369785] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] result = hub.switch() [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self.greenlet.switch() [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] result = function(*args, **kwargs) [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return func(*args, **kwargs) [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise e [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] nwinfo = self.network_api.allocate_for_instance( [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.370209] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] created_port_ids = self._update_ports_for_instance( [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] with excutils.save_and_reraise_exception(): [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.force_reraise() [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise self.value [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] updated_port = self._update_port( [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] _ensure_no_port_binding_failure(port) [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.370624] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise exception.PortBindingFailed(port_id=port['id']) [ 582.371189] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 582.371189] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] [ 582.371189] env[62066]: INFO nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Terminating instance [ 582.371757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.450555] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.547823] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.752401] env[62066]: DEBUG nova.network.neutron [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.889800] env[62066]: DEBUG nova.network.neutron [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.050448] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-fc176994-95b2-44a3-becf-6c224476851a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.051902] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.052214] env[62066]: DEBUG nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.052948] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.069655] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.380048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa3ad10-d331-4336-a2ce-dade479a1b7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.389839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74921226-e343-427a-846b-0e5fa016e1cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.397024] env[62066]: DEBUG oslo_concurrency.lockutils [req-de9a6a17-08bc-4962-b456-167c2d0797c0 req-6ae131c1-422a-4845-b17a-a9f4e1385ccd service nova] Releasing lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.397024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquired lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.397024] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.427805] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962e6b6d-af55-4536-9445-f405122ade68 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.437654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10528b63-eb42-4576-9f40-3317618f5056 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.454336] env[62066]: DEBUG nova.compute.provider_tree [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.574634] env[62066]: DEBUG nova.network.neutron [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.918126] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.959890] env[62066]: DEBUG nova.scheduler.client.report [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.010659] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.077876] env[62066]: INFO nova.compute.manager [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: fc176994-95b2-44a3-becf-6c224476851a] Took 1.03 seconds to deallocate network for instance. [ 584.302353] env[62066]: DEBUG nova.compute.manager [req-b036959d-a614-4a63-96c0-d2c5edc0be88 req-2f8381d4-ec3e-484e-ad28-07f290fcd3d4 service nova] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Received event network-vif-deleted-803ac5bc-080a-4690-b757-02374a841ea3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.328836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquiring lock "ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.329132] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.466384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.470347] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.471521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.464s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.517329] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Releasing lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.518331] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.518331] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.518331] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5ebcb96-c852-4462-8f1e-fdc6aee2abb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.529846] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f516e90-d0ed-462b-b6f3-f6bc720ac58c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.554363] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf3dc6c9-d7bf-4319-9420-e3634b9fca11 could not be found. [ 584.554556] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.554737] env[62066]: INFO nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.554970] env[62066]: DEBUG oslo.service.loopingcall [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.555202] env[62066]: DEBUG nova.compute.manager [-] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.555294] env[62066]: DEBUG nova.network.neutron [-] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 584.590075] env[62066]: DEBUG nova.network.neutron [-] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.977689] env[62066]: DEBUG nova.compute.utils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 584.987737] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 584.987939] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 585.054173] env[62066]: DEBUG nova.policy [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c575e8611164519b3fa7774ffdfc892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42be1b4424d540849df5f695a386f49b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.101930] env[62066]: DEBUG nova.network.neutron [-] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.144957] env[62066]: INFO nova.scheduler.client.report [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Deleted allocations for instance fc176994-95b2-44a3-becf-6c224476851a [ 585.467599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd0a022-4c09-4978-a51c-44ec59620f57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.475876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8482e27a-4155-4e33-8489-12ef47e87798 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.508183] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 585.512720] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e08d5e-c1d8-452f-8478-d44c76825716 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.520069] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7992d255-2681-45fc-b706-3a2d8bbf993f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.535675] env[62066]: DEBUG nova.compute.provider_tree [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.605383] env[62066]: INFO nova.compute.manager [-] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Took 1.05 seconds to deallocate network for instance. [ 585.610286] env[62066]: DEBUG nova.compute.claims [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.610286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.654275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d5f3f642-1386-4034-b5e5-c3da609294c8 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "fc176994-95b2-44a3-becf-6c224476851a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.730s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.727790] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Successfully created port: 1e296eab-a517-4d4f-b471-1a3042b5a5bd {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.038934] env[62066]: DEBUG nova.scheduler.client.report [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.157848] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.518158] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 586.546470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.075s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.548628] env[62066]: ERROR nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Traceback (most recent call last): [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.driver.spawn(context, instance, image_meta, [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] vm_ref = self.build_virtual_machine(instance, [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.548628] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] for vif in network_info: [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return self._sync_wrapper(fn, *args, **kwargs) [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.wait() [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self[:] = self._gt.wait() [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return self._exit_event.wait() [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] current.throw(*self._exc) [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.549025] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] result = function(*args, **kwargs) [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] return func(*args, **kwargs) [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise e [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] nwinfo = self.network_api.allocate_for_instance( [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] created_port_ids = self._update_ports_for_instance( [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] with excutils.save_and_reraise_exception(): [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] self.force_reraise() [ 586.549711] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise self.value [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] updated_port = self._update_port( [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] _ensure_no_port_binding_failure(port) [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] raise exception.PortBindingFailed(port_id=port['id']) [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] nova.exception.PortBindingFailed: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. [ 586.550148] env[62066]: ERROR nova.compute.manager [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] [ 586.550148] env[62066]: DEBUG nova.compute.utils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 586.551116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.046s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.555433] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Build of instance 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd was re-scheduled: Binding failed for port 2c3f593c-19f7-4f71-aeb2-ade6950ddd19, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.555834] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.556062] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquiring lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.556666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Acquired lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.556666] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.560681] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.560785] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.560923] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.563975] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.564164] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.564320] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.564537] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.564692] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.564849] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.565015] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.565195] env[62066]: DEBUG nova.virt.hardware [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.570111] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3c49f6-5680-4ebb-a59b-0987a7152169 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.580453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb7d7c2-0205-4d1e-ae27-9117d93338c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.679610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.944661] env[62066]: DEBUG nova.compute.manager [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Received event network-changed-1e296eab-a517-4d4f-b471-1a3042b5a5bd {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.944834] env[62066]: DEBUG nova.compute.manager [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Refreshing instance network info cache due to event network-changed-1e296eab-a517-4d4f-b471-1a3042b5a5bd. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.946666] env[62066]: DEBUG oslo_concurrency.lockutils [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] Acquiring lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.946666] env[62066]: DEBUG oslo_concurrency.lockutils [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] Acquired lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.946666] env[62066]: DEBUG nova.network.neutron [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Refreshing network info cache for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 587.101289] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.177490] env[62066]: ERROR nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 587.177490] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.177490] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.177490] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.177490] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.177490] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.177490] env[62066]: ERROR nova.compute.manager raise self.value [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.177490] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.177490] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.177490] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.178593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.178593] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.178593] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 587.178593] env[62066]: ERROR nova.compute.manager [ 587.178593] env[62066]: Traceback (most recent call last): [ 587.178593] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.178593] env[62066]: listener.cb(fileno) [ 587.178593] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.178593] env[62066]: result = function(*args, **kwargs) [ 587.178593] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.178593] env[62066]: return func(*args, **kwargs) [ 587.178593] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.178593] env[62066]: raise e [ 587.178593] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.178593] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 587.178593] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.178593] env[62066]: created_port_ids = self._update_ports_for_instance( [ 587.178593] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.178593] env[62066]: with excutils.save_and_reraise_exception(): [ 587.178593] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.178593] env[62066]: self.force_reraise() [ 587.178593] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.178593] env[62066]: raise self.value [ 587.178593] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.178593] env[62066]: updated_port = self._update_port( [ 587.178593] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.178593] env[62066]: _ensure_no_port_binding_failure(port) [ 587.178593] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.178593] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.180043] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 587.180043] env[62066]: Removing descriptor: 18 [ 587.180043] env[62066]: ERROR nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Traceback (most recent call last): [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] yield resources [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.driver.spawn(context, instance, image_meta, [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.180043] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] vm_ref = self.build_virtual_machine(instance, [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] for vif in network_info: [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self._sync_wrapper(fn, *args, **kwargs) [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.wait() [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self[:] = self._gt.wait() [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self._exit_event.wait() [ 587.180834] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] result = hub.switch() [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self.greenlet.switch() [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] result = function(*args, **kwargs) [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return func(*args, **kwargs) [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise e [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] nwinfo = self.network_api.allocate_for_instance( [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.181530] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] created_port_ids = self._update_ports_for_instance( [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] with excutils.save_and_reraise_exception(): [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.force_reraise() [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise self.value [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] updated_port = self._update_port( [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] _ensure_no_port_binding_failure(port) [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.182114] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise exception.PortBindingFailed(port_id=port['id']) [ 587.182588] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 587.182588] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] [ 587.182588] env[62066]: INFO nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Terminating instance [ 587.182588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.221025] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.481216] env[62066]: DEBUG nova.network.neutron [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.530343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e165f178-67a4-4fbf-b58b-9e04362ec8af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.539271] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41516be0-aceb-451c-92f9-b0ac87b344dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.571330] env[62066]: DEBUG nova.network.neutron [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.572988] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca34154d-6cbf-41c7-8291-277c452ecf0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.583951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58475d0e-5882-4355-a083-b2d9e442c858 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.599667] env[62066]: DEBUG nova.compute.provider_tree [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.727021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Releasing lock "refresh_cache-212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.727272] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.727448] env[62066]: DEBUG nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.727692] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.747294] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.023224] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.023733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.077208] env[62066]: DEBUG oslo_concurrency.lockutils [req-93b4b67f-c6cb-4a93-9c44-208b8190d101 req-3c71c1ae-449e-4cae-838b-fe7124b3a0d0 service nova] Releasing lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.077672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquired lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.077862] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.103983] env[62066]: DEBUG nova.scheduler.client.report [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.251343] env[62066]: DEBUG nova.network.neutron [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.599805] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.611255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.611886] env[62066]: ERROR nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Traceback (most recent call last): [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.driver.spawn(context, instance, image_meta, [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] vm_ref = self.build_virtual_machine(instance, [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.611886] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] for vif in network_info: [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return self._sync_wrapper(fn, *args, **kwargs) [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.wait() [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self[:] = self._gt.wait() [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return self._exit_event.wait() [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] current.throw(*self._exc) [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.612326] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] result = function(*args, **kwargs) [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] return func(*args, **kwargs) [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise e [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] nwinfo = self.network_api.allocate_for_instance( [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] created_port_ids = self._update_ports_for_instance( [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] with excutils.save_and_reraise_exception(): [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] self.force_reraise() [ 588.612739] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise self.value [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] updated_port = self._update_port( [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] _ensure_no_port_binding_failure(port) [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] raise exception.PortBindingFailed(port_id=port['id']) [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] nova.exception.PortBindingFailed: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. [ 588.613202] env[62066]: ERROR nova.compute.manager [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] [ 588.613202] env[62066]: DEBUG nova.compute.utils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 588.613826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.498s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.616727] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Build of instance bdf32142-16bd-429a-ac8b-e0aa0c47b5d0 was re-scheduled: Binding failed for port 6894146e-8eb2-4de3-8e71-cd02b26a5c14, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 588.617176] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 588.617404] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquiring lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.617552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Acquired lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.617707] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.683717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.684310] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.698462] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.756052] env[62066]: INFO nova.compute.manager [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] [instance: 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd] Took 1.03 seconds to deallocate network for instance. [ 588.968523] env[62066]: DEBUG nova.compute.manager [req-1264987f-78c1-4e81-8009-720dec050bc8 req-36090f7f-c626-43a1-b233-6b956d1e3ffb service nova] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Received event network-vif-deleted-1e296eab-a517-4d4f-b471-1a3042b5a5bd {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.139062] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.200757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Releasing lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.201199] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.201393] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.201709] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d8a00d1-e493-4fd4-a25f-262b178ba91f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.213035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee63730-3d5c-46d4-a9e7-66a6b6812326 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.225865] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.244073] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 30f7baa5-0288-4ef3-9d1c-d7143868b3a4 could not be found. [ 589.244302] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.244480] env[62066]: INFO nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.244712] env[62066]: DEBUG oslo.service.loopingcall [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.244928] env[62066]: DEBUG nova.compute.manager [-] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.245032] env[62066]: DEBUG nova.network.neutron [-] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.259768] env[62066]: DEBUG nova.network.neutron [-] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.478281] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3914319c-089a-4e93-a873-53fab1300355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.486403] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139cb483-e11a-46fd-8bbd-784c82ee777f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.516470] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5b0f7f-df56-47a6-9f5e-1e817d72e2ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.524398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2faee298-d17d-4ff9-818d-89824ec1c0ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.537562] env[62066]: DEBUG nova.compute.provider_tree [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.729191] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Releasing lock "refresh_cache-bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.729542] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 589.730027] env[62066]: DEBUG nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.730027] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.746031] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.764717] env[62066]: DEBUG nova.network.neutron [-] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.787060] env[62066]: INFO nova.scheduler.client.report [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Deleted allocations for instance 212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd [ 590.040686] env[62066]: DEBUG nova.scheduler.client.report [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.249662] env[62066]: DEBUG nova.network.neutron [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.266565] env[62066]: INFO nova.compute.manager [-] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Took 1.02 seconds to deallocate network for instance. [ 590.268859] env[62066]: DEBUG nova.compute.claims [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.269055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.295841] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4bc1e624-bfc9-40ee-bccc-5f01a6bccd15 tempest-ServersWithSpecificFlavorTestJSON-2082978471 tempest-ServersWithSpecificFlavorTestJSON-2082978471-project-member] Lock "212cb5fc-f7ca-48cb-ac72-a74d4b17e9cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.973s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.546348] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.547345] env[62066]: ERROR nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Traceback (most recent call last): [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.driver.spawn(context, instance, image_meta, [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] vm_ref = self.build_virtual_machine(instance, [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.547345] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] for vif in network_info: [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return self._sync_wrapper(fn, *args, **kwargs) [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.wait() [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self[:] = self._gt.wait() [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return self._exit_event.wait() [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] current.throw(*self._exc) [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.548516] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] result = function(*args, **kwargs) [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] return func(*args, **kwargs) [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise e [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] nwinfo = self.network_api.allocate_for_instance( [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] created_port_ids = self._update_ports_for_instance( [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] with excutils.save_and_reraise_exception(): [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] self.force_reraise() [ 590.551637] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise self.value [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] updated_port = self._update_port( [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] _ensure_no_port_binding_failure(port) [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] raise exception.PortBindingFailed(port_id=port['id']) [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] nova.exception.PortBindingFailed: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. [ 590.552131] env[62066]: ERROR nova.compute.manager [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] [ 590.552131] env[62066]: DEBUG nova.compute.utils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 590.552489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.895s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.557165] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Build of instance 92ca97dd-cd7c-4612-85d9-a05f5a785d6f was re-scheduled: Binding failed for port 52508203-efc9-4b34-b970-4debe6b97ded, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 590.558923] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 590.558923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquiring lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.558923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Acquired lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.558923] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.752267] env[62066]: INFO nova.compute.manager [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] [instance: bdf32142-16bd-429a-ac8b-e0aa0c47b5d0] Took 1.02 seconds to deallocate network for instance. [ 590.803234] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.083198] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.205468] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.320620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.454616] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7475ec59-50f0-4126-a2d1-9250fc551afe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.463304] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296a462a-d5cf-41d0-8898-c51168fae8e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.498690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11948a7-4bc4-4d40-b20b-9932aad94175 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.507423] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f5a683-2f67-46b3-aec8-7965897ad923 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.523832] env[62066]: DEBUG nova.compute.provider_tree [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.708295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Releasing lock "refresh_cache-92ca97dd-cd7c-4612-85d9-a05f5a785d6f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.708659] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 591.708916] env[62066]: DEBUG nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.709162] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.723703] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.784774] env[62066]: INFO nova.scheduler.client.report [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Deleted allocations for instance bdf32142-16bd-429a-ac8b-e0aa0c47b5d0 [ 592.027990] env[62066]: DEBUG nova.scheduler.client.report [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.226877] env[62066]: DEBUG nova.network.neutron [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.301371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02297bc4-268f-4fee-8c39-7998c04e47ae tempest-ServerMetadataNegativeTestJSON-1267124696 tempest-ServerMetadataNegativeTestJSON-1267124696-project-member] Lock "bdf32142-16bd-429a-ac8b-e0aa0c47b5d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.577s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.540946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.541582] env[62066]: ERROR nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Traceback (most recent call last): [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.driver.spawn(context, instance, image_meta, [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] vm_ref = self.build_virtual_machine(instance, [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.541582] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] for vif in network_info: [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return self._sync_wrapper(fn, *args, **kwargs) [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.wait() [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self[:] = self._gt.wait() [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return self._exit_event.wait() [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] current.throw(*self._exc) [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.541980] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] result = function(*args, **kwargs) [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] return func(*args, **kwargs) [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise e [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] nwinfo = self.network_api.allocate_for_instance( [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] created_port_ids = self._update_ports_for_instance( [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] with excutils.save_and_reraise_exception(): [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] self.force_reraise() [ 592.542377] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise self.value [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] updated_port = self._update_port( [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] _ensure_no_port_binding_failure(port) [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] raise exception.PortBindingFailed(port_id=port['id']) [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] nova.exception.PortBindingFailed: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. [ 592.542740] env[62066]: ERROR nova.compute.manager [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] [ 592.542740] env[62066]: DEBUG nova.compute.utils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 592.543707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.276s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.545032] env[62066]: INFO nova.compute.claims [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.552307] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Build of instance dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b was re-scheduled: Binding failed for port c2cc0772-f39d-456a-8d4a-cc2b1f5bd1a4, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.552307] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.552307] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquiring lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.552307] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Acquired lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.552551] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.730956] env[62066]: INFO nova.compute.manager [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] [instance: 92ca97dd-cd7c-4612-85d9-a05f5a785d6f] Took 1.02 seconds to deallocate network for instance. [ 592.803760] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.075840] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.196465] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.330347] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.703025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Releasing lock "refresh_cache-dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.703025] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.703025] env[62066]: DEBUG nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.703025] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.725876] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.767741] env[62066]: INFO nova.scheduler.client.report [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Deleted allocations for instance 92ca97dd-cd7c-4612-85d9-a05f5a785d6f [ 593.992152] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4725871d-7d84-4a32-a866-d03f0f1f9c3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.004712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e19660-bf39-4557-af24-5fe591c72267 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.054701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ba43c5-6a15-4ce7-a6eb-ca1d4f7e73e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.064092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c1e7e2-12f5-4ab4-8ef0-2685dd511260 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.079709] env[62066]: DEBUG nova.compute.provider_tree [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.232037] env[62066]: DEBUG nova.network.neutron [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.282720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a00d39cd-1e4b-4dd6-b982-4badd77e42aa tempest-ImagesOneServerTestJSON-1453867304 tempest-ImagesOneServerTestJSON-1453867304-project-member] Lock "92ca97dd-cd7c-4612-85d9-a05f5a785d6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.341s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.584581] env[62066]: DEBUG nova.scheduler.client.report [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.736958] env[62066]: INFO nova.compute.manager [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] [instance: dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b] Took 1.04 seconds to deallocate network for instance. [ 594.785333] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.093621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.094348] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.097305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.500s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.193685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.193904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.319750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.519279] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "2f186d0a-91a0-4dc2-83bd-511099445af7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.519640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.602688] env[62066]: DEBUG nova.compute.utils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 595.606674] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.606843] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.697806] env[62066]: DEBUG nova.policy [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dc6a6719c2446bd8f1893c9aa4a8a74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f7c5676e74d4a699a84310375e3643e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.779764] env[62066]: INFO nova.scheduler.client.report [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Deleted allocations for instance dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b [ 596.037173] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a384a9b-7561-4395-976b-41b058ed49b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.046638] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d61bcb5-df51-4235-8d39-1c62abdedac4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.082157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8625351-cc92-4dcf-bc16-5f2b04894a38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.090871] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2adcb5d-1853-4217-9d7f-bd887d6c5735 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.105080] env[62066]: DEBUG nova.compute.provider_tree [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.110577] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.296285] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4740ac72-76a6-4bee-a6a3-4e0066dbf13e tempest-DeleteServersAdminTestJSON-370916226 tempest-DeleteServersAdminTestJSON-370916226-project-member] Lock "dfcf6263-d4b9-4d4b-aa5b-8eeb31c2899b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.853s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.360293] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Successfully created port: ec358e5f-1fb3-406a-98ca-75864e8177e0 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.607553] env[62066]: DEBUG nova.scheduler.client.report [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.799284] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.122308] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.122308] env[62066]: ERROR nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Traceback (most recent call last): [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.driver.spawn(context, instance, image_meta, [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.122308] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] vm_ref = self.build_virtual_machine(instance, [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] for vif in network_info: [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self._sync_wrapper(fn, *args, **kwargs) [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.wait() [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self[:] = self._gt.wait() [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self._exit_event.wait() [ 597.122725] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] result = hub.switch() [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return self.greenlet.switch() [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] result = function(*args, **kwargs) [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] return func(*args, **kwargs) [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise e [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] nwinfo = self.network_api.allocate_for_instance( [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.122991] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] created_port_ids = self._update_ports_for_instance( [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] with excutils.save_and_reraise_exception(): [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] self.force_reraise() [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise self.value [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] updated_port = self._update_port( [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] _ensure_no_port_binding_failure(port) [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.123288] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] raise exception.PortBindingFailed(port_id=port['id']) [ 597.123522] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] nova.exception.PortBindingFailed: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. [ 597.123522] env[62066]: ERROR nova.compute.manager [instance: b3d0de63-fe22-4bc7-9602-509054dca586] [ 597.124479] env[62066]: DEBUG nova.compute.utils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 597.126084] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Build of instance b3d0de63-fe22-4bc7-9602-509054dca586 was re-scheduled: Binding failed for port c4e9385c-f0d1-4cb2-a486-80a4186def11, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.126520] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.127033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquiring lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.127033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Acquired lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.127103] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.128703] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.132455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.122s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.133960] env[62066]: INFO nova.compute.claims [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.174925] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.175716] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.175716] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.175716] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.175716] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.175832] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.176036] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.178210] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.178210] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.178210] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.178210] env[62066]: DEBUG nova.virt.hardware [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.182225] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bd6a22-4858-4f05-a621-4c2850fbcdbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.192918] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc9eeea-9809-443f-972c-241ee5ca90ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.247900] env[62066]: DEBUG nova.compute.manager [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Received event network-changed-ec358e5f-1fb3-406a-98ca-75864e8177e0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.248332] env[62066]: DEBUG nova.compute.manager [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Refreshing instance network info cache due to event network-changed-ec358e5f-1fb3-406a-98ca-75864e8177e0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.248555] env[62066]: DEBUG oslo_concurrency.lockutils [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] Acquiring lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.248876] env[62066]: DEBUG oslo_concurrency.lockutils [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] Acquired lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.248976] env[62066]: DEBUG nova.network.neutron [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Refreshing network info cache for port ec358e5f-1fb3-406a-98ca-75864e8177e0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.329835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.356099] env[62066]: ERROR nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 597.356099] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.356099] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.356099] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.356099] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.356099] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.356099] env[62066]: ERROR nova.compute.manager raise self.value [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.356099] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.356099] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.356099] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.356521] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.356521] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.356521] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 597.356521] env[62066]: ERROR nova.compute.manager [ 597.356521] env[62066]: Traceback (most recent call last): [ 597.356521] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.356521] env[62066]: listener.cb(fileno) [ 597.356521] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.356521] env[62066]: result = function(*args, **kwargs) [ 597.356521] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.356521] env[62066]: return func(*args, **kwargs) [ 597.356521] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.356521] env[62066]: raise e [ 597.356521] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.356521] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 597.356521] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.356521] env[62066]: created_port_ids = self._update_ports_for_instance( [ 597.356521] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.356521] env[62066]: with excutils.save_and_reraise_exception(): [ 597.356521] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.356521] env[62066]: self.force_reraise() [ 597.356521] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.356521] env[62066]: raise self.value [ 597.356521] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.356521] env[62066]: updated_port = self._update_port( [ 597.356521] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.356521] env[62066]: _ensure_no_port_binding_failure(port) [ 597.356521] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.356521] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.357212] env[62066]: nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 597.357212] env[62066]: Removing descriptor: 19 [ 597.357212] env[62066]: ERROR nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] Traceback (most recent call last): [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] yield resources [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.driver.spawn(context, instance, image_meta, [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.357212] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] vm_ref = self.build_virtual_machine(instance, [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] for vif in network_info: [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self._sync_wrapper(fn, *args, **kwargs) [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.wait() [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self[:] = self._gt.wait() [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self._exit_event.wait() [ 597.357585] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] result = hub.switch() [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self.greenlet.switch() [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] result = function(*args, **kwargs) [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return func(*args, **kwargs) [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise e [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] nwinfo = self.network_api.allocate_for_instance( [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.357908] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] created_port_ids = self._update_ports_for_instance( [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] with excutils.save_and_reraise_exception(): [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.force_reraise() [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise self.value [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] updated_port = self._update_port( [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] _ensure_no_port_binding_failure(port) [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.358295] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise exception.PortBindingFailed(port_id=port['id']) [ 597.358589] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 597.358589] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] [ 597.358589] env[62066]: INFO nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Terminating instance [ 597.359780] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquiring lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.652122] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.701245] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.769554] env[62066]: DEBUG nova.network.neutron [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.819042] env[62066]: DEBUG nova.network.neutron [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.203892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Releasing lock "refresh_cache-b3d0de63-fe22-4bc7-9602-509054dca586" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.205294] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.205294] env[62066]: DEBUG nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.205294] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.225674] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.322782] env[62066]: DEBUG oslo_concurrency.lockutils [req-81eb39c8-5194-481f-95e0-17ee478a1ae7 req-4862a6b9-613a-49f5-ba90-80772197b17b service nova] Releasing lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.323700] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquired lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.323700] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.531589] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8bc46d9-46a7-439e-9b6b-45de2424718a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.540075] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc877ba-f64e-4f45-837f-5a0157b2dd2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.572887] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd941a61-3d45-4f9a-a2ef-6b3c0d3a394b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.582436] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7678d69d-a61b-4fb3-b7e7-f999ada3a0ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.596989] env[62066]: DEBUG nova.compute.provider_tree [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.728951] env[62066]: DEBUG nova.network.neutron [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.843949] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.904234] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.103026] env[62066]: DEBUG nova.scheduler.client.report [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.234046] env[62066]: INFO nova.compute.manager [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] [instance: b3d0de63-fe22-4bc7-9602-509054dca586] Took 1.03 seconds to deallocate network for instance. [ 599.282168] env[62066]: DEBUG nova.compute.manager [req-c92cb1a5-d7d1-4b53-8ac9-3c555b463b4d req-b88f072a-821f-412a-97ff-f317cf1dcbec service nova] [instance: de135f73-53d2-4471-a0ba-9204af726690] Received event network-vif-deleted-ec358e5f-1fb3-406a-98ca-75864e8177e0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.407469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Releasing lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.408046] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.408258] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.408573] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97abd2b3-a31f-4d49-9c8a-7e1a87ebfbb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.419373] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c6e54-58da-4c0a-8f1e-7fee3b3092f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.442130] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de135f73-53d2-4471-a0ba-9204af726690 could not be found. [ 599.442409] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.442601] env[62066]: INFO nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Took 0.03 seconds to destroy the instance on the hypervisor. [ 599.442836] env[62066]: DEBUG oslo.service.loopingcall [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.443062] env[62066]: DEBUG nova.compute.manager [-] [instance: de135f73-53d2-4471-a0ba-9204af726690] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.443159] env[62066]: DEBUG nova.network.neutron [-] [instance: de135f73-53d2-4471-a0ba-9204af726690] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.458432] env[62066]: DEBUG nova.network.neutron [-] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.605756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.606673] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.609299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.376s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.609961] env[62066]: DEBUG nova.objects.instance [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lazy-loading 'resources' on Instance uuid fa73770f-784a-44c3-ba65-e7176180760b {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 599.963225] env[62066]: DEBUG nova.network.neutron [-] [instance: de135f73-53d2-4471-a0ba-9204af726690] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.114125] env[62066]: DEBUG nova.compute.utils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 600.122914] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.122914] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.183253] env[62066]: DEBUG nova.policy [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7407a43b92d47d6be52b56a2fd156a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0205036823fe48f5b8b91f7d287b2511', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.272782] env[62066]: INFO nova.scheduler.client.report [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Deleted allocations for instance b3d0de63-fe22-4bc7-9602-509054dca586 [ 600.465975] env[62066]: INFO nova.compute.manager [-] [instance: de135f73-53d2-4471-a0ba-9204af726690] Took 1.02 seconds to deallocate network for instance. [ 600.467923] env[62066]: DEBUG nova.compute.claims [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.468134] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.547243] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a37fa6-208a-4ea8-853a-beb0efeb4da5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.558662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0271d56-ce3b-4a34-9850-c45f2dcd9703 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.602500] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccc9a97-c29e-4e33-a97a-cf5d8bb7dd71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.612025] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f175cea-f97d-41e3-9463-eb6e128c440f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.627409] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.630140] env[62066]: DEBUG nova.compute.provider_tree [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.631757] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Successfully created port: 309971aa-38f4-4b36-b889-eb6ee86b8200 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.794436] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eddda0d4-4067-4096-95cf-29510c6aff8a tempest-ServerDiagnosticsTest-115863366 tempest-ServerDiagnosticsTest-115863366-project-member] Lock "b3d0de63-fe22-4bc7-9602-509054dca586" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.201s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.139321] env[62066]: DEBUG nova.scheduler.client.report [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.297228] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.644902] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.650452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.652048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.043s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.680671] env[62066]: INFO nova.scheduler.client.report [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Deleted allocations for instance fa73770f-784a-44c3-ba65-e7176180760b [ 601.685094] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.685368] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.685588] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.685783] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.685923] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.690299] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.691023] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.691023] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.691023] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.691169] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.691338] env[62066]: DEBUG nova.virt.hardware [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.692272] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20339e0-ec20-4dcd-8765-786bc1b5d567 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.708080] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fa4c35-7265-46a7-bc39-a716efb992a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.830194] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.189019] env[62066]: DEBUG nova.compute.manager [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Received event network-changed-309971aa-38f4-4b36-b889-eb6ee86b8200 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.189019] env[62066]: DEBUG nova.compute.manager [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Refreshing instance network info cache due to event network-changed-309971aa-38f4-4b36-b889-eb6ee86b8200. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.189019] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] Acquiring lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.189019] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] Acquired lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.189019] env[62066]: DEBUG nova.network.neutron [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Refreshing network info cache for port 309971aa-38f4-4b36-b889-eb6ee86b8200 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 602.212251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ec53bfe-a3bd-426d-8446-95b2580d8f10 tempest-ServerDiagnosticsV248Test-923191108 tempest-ServerDiagnosticsV248Test-923191108-project-member] Lock "fa73770f-784a-44c3-ba65-e7176180760b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.759s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.362231] env[62066]: ERROR nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 602.362231] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.362231] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.362231] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.362231] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.362231] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.362231] env[62066]: ERROR nova.compute.manager raise self.value [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.362231] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.362231] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.362231] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.362880] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.362880] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.362880] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 602.362880] env[62066]: ERROR nova.compute.manager [ 602.362880] env[62066]: Traceback (most recent call last): [ 602.362880] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.362880] env[62066]: listener.cb(fileno) [ 602.362880] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.362880] env[62066]: result = function(*args, **kwargs) [ 602.362880] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.362880] env[62066]: return func(*args, **kwargs) [ 602.362880] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.362880] env[62066]: raise e [ 602.362880] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.362880] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 602.362880] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.362880] env[62066]: created_port_ids = self._update_ports_for_instance( [ 602.362880] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.362880] env[62066]: with excutils.save_and_reraise_exception(): [ 602.362880] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.362880] env[62066]: self.force_reraise() [ 602.362880] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.362880] env[62066]: raise self.value [ 602.362880] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.362880] env[62066]: updated_port = self._update_port( [ 602.362880] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.362880] env[62066]: _ensure_no_port_binding_failure(port) [ 602.362880] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.362880] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.363518] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 602.363518] env[62066]: Removing descriptor: 19 [ 602.363518] env[62066]: ERROR nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Traceback (most recent call last): [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] yield resources [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.driver.spawn(context, instance, image_meta, [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.363518] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] vm_ref = self.build_virtual_machine(instance, [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] for vif in network_info: [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self._sync_wrapper(fn, *args, **kwargs) [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.wait() [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self[:] = self._gt.wait() [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self._exit_event.wait() [ 602.363772] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] result = hub.switch() [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self.greenlet.switch() [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] result = function(*args, **kwargs) [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return func(*args, **kwargs) [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise e [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] nwinfo = self.network_api.allocate_for_instance( [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.364080] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] created_port_ids = self._update_ports_for_instance( [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] with excutils.save_and_reraise_exception(): [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.force_reraise() [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise self.value [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] updated_port = self._update_port( [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] _ensure_no_port_binding_failure(port) [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.364369] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise exception.PortBindingFailed(port_id=port['id']) [ 602.364625] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 602.364625] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] [ 602.364625] env[62066]: INFO nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Terminating instance [ 602.367064] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquiring lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.626349] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773fb62e-67f8-4f29-aabb-92a588c547a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.641028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd0ee36-bb7e-4ffe-b05d-1486f55622fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.678925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb42eb2-afce-4aaf-b800-0b859d3903d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.688291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d6b13d-581f-4411-8fc0-5daf1970931b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.708542] env[62066]: DEBUG nova.compute.provider_tree [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.773051] env[62066]: DEBUG nova.network.neutron [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.999689] env[62066]: DEBUG nova.network.neutron [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.212280] env[62066]: DEBUG nova.scheduler.client.report [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.507044] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee5f2759-4362-4d97-a16d-f13b136940b7 req-60da521d-9e0a-4d25-b734-4c3ec6a561f2 service nova] Releasing lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.507044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquired lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.507302] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.717111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.066s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.721126] env[62066]: ERROR nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Traceback (most recent call last): [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.driver.spawn(context, instance, image_meta, [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] vm_ref = self.build_virtual_machine(instance, [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.721126] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] for vif in network_info: [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self._sync_wrapper(fn, *args, **kwargs) [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.wait() [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self[:] = self._gt.wait() [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self._exit_event.wait() [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] result = hub.switch() [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.721492] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return self.greenlet.switch() [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] result = function(*args, **kwargs) [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] return func(*args, **kwargs) [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise e [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] nwinfo = self.network_api.allocate_for_instance( [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] created_port_ids = self._update_ports_for_instance( [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] with excutils.save_and_reraise_exception(): [ 603.721826] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] self.force_reraise() [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise self.value [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] updated_port = self._update_port( [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] _ensure_no_port_binding_failure(port) [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] raise exception.PortBindingFailed(port_id=port['id']) [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] nova.exception.PortBindingFailed: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. [ 603.722218] env[62066]: ERROR nova.compute.manager [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] [ 603.722522] env[62066]: DEBUG nova.compute.utils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 603.722522] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.040s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.722522] env[62066]: INFO nova.compute.claims [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.725684] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Build of instance cf3dc6c9-d7bf-4319-9420-e3634b9fca11 was re-scheduled: Binding failed for port 803ac5bc-080a-4690-b757-02374a841ea3, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 603.726150] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 603.726373] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.726513] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquired lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.726665] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.043043] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.192595] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.257020] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.526490] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.700022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Releasing lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.700022] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.700022] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 604.700022] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-744b6e17-f10b-4fe0-9552-cec9720a996b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.711519] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e2b96d-2f75-44e6-91ce-140e981e2d3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.741936] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8bad79f-9bf7-4ef2-b681-ca8901e878fd could not be found. [ 604.742076] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.742214] env[62066]: INFO nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.742676] env[62066]: DEBUG oslo.service.loopingcall [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.743856] env[62066]: DEBUG nova.compute.manager [-] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.743940] env[62066]: DEBUG nova.network.neutron [-] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.767573] env[62066]: DEBUG nova.network.neutron [-] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.850441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.850734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.895763] env[62066]: DEBUG nova.compute.manager [req-0e4b5866-74ed-427b-ac4b-069653f8da58 req-846dcb62-8456-444a-be91-387fdea714ce service nova] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Received event network-vif-deleted-309971aa-38f4-4b36-b889-eb6ee86b8200 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.032951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Releasing lock "refresh_cache-cf3dc6c9-d7bf-4319-9420-e3634b9fca11" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.033236] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.036788] env[62066]: DEBUG nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.036788] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.063479] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.232261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "d50f0505-66f4-412f-9744-25e0ea96277a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.232912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.265379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32d0596-6c2c-4bb3-86bf-a220be901f0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.269914] env[62066]: DEBUG nova.network.neutron [-] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.276563] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0f43b3-8324-41c8-bb3f-440dbf7d0bd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.312790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e77abaa-5dfe-4abd-98fe-f228b6a2cadc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.321655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ee46ab-938a-4fbf-bc20-5b111c235658 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.337285] env[62066]: DEBUG nova.compute.provider_tree [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.565212] env[62066]: DEBUG nova.network.neutron [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.429021] env[62066]: INFO nova.compute.manager [-] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Took 1.68 seconds to deallocate network for instance. [ 606.430351] env[62066]: DEBUG nova.scheduler.client.report [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.433581] env[62066]: INFO nova.compute.manager [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: cf3dc6c9-d7bf-4319-9420-e3634b9fca11] Took 1.40 seconds to deallocate network for instance. [ 606.438967] env[62066]: DEBUG nova.compute.claims [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.438967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.561100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "a8dd7483-0588-4f60-9504-20de799e69f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.561527] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.938157] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.218s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.938719] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.942222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.673s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.453361] env[62066]: DEBUG nova.compute.utils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 607.456059] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.456059] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.484738] env[62066]: INFO nova.scheduler.client.report [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Deleted allocations for instance cf3dc6c9-d7bf-4319-9420-e3634b9fca11 [ 607.511256] env[62066]: DEBUG nova.policy [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81f6be3be50146bd9a050d9b34ac2dfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8ed42d4f0cb40e2aea29a7b9a14a6f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.823534] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Successfully created port: 534dfc23-9752-49e0-a34a-c21f75401d5e {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.953827] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d61cacd-5317-40e1-9a1b-255908a40af7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.958534] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.965324] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c336c1d-1895-43c2-8643-9f0a78cada58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.000595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28992414-be92-42e7-a8ba-0b2016c94494 tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "cf3dc6c9-d7bf-4319-9420-e3634b9fca11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.317s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.006032] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52bf72a-a0da-41f6-a4e3-0eb0f7b5f9bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.011526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cf8b0d-6010-439b-ae55-0c27bc346e89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.026357] env[62066]: DEBUG nova.compute.provider_tree [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.083572] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.083805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.506886] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.531768] env[62066]: DEBUG nova.scheduler.client.report [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.588744] env[62066]: DEBUG nova.compute.manager [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Received event network-changed-534dfc23-9752-49e0-a34a-c21f75401d5e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.588944] env[62066]: DEBUG nova.compute.manager [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Refreshing instance network info cache due to event network-changed-534dfc23-9752-49e0-a34a-c21f75401d5e. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.589212] env[62066]: DEBUG oslo_concurrency.lockutils [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] Acquiring lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.589360] env[62066]: DEBUG oslo_concurrency.lockutils [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] Acquired lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.589561] env[62066]: DEBUG nova.network.neutron [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Refreshing network info cache for port 534dfc23-9752-49e0-a34a-c21f75401d5e {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 608.736674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.736899] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.759986] env[62066]: ERROR nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 608.759986] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.759986] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.759986] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.759986] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.759986] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.759986] env[62066]: ERROR nova.compute.manager raise self.value [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.759986] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.759986] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.759986] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.760524] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.760524] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.760524] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 608.760524] env[62066]: ERROR nova.compute.manager [ 608.760524] env[62066]: Traceback (most recent call last): [ 608.760524] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.760524] env[62066]: listener.cb(fileno) [ 608.760524] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.760524] env[62066]: result = function(*args, **kwargs) [ 608.760524] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.760524] env[62066]: return func(*args, **kwargs) [ 608.760524] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.760524] env[62066]: raise e [ 608.760524] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.760524] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 608.760524] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.760524] env[62066]: created_port_ids = self._update_ports_for_instance( [ 608.760524] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.760524] env[62066]: with excutils.save_and_reraise_exception(): [ 608.760524] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.760524] env[62066]: self.force_reraise() [ 608.760524] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.760524] env[62066]: raise self.value [ 608.760524] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.760524] env[62066]: updated_port = self._update_port( [ 608.760524] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.760524] env[62066]: _ensure_no_port_binding_failure(port) [ 608.760524] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.760524] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.761173] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 608.761173] env[62066]: Removing descriptor: 19 [ 608.976031] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.001447] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.001748] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.001924] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.002117] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.002719] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.002719] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.002719] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.002864] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.003026] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.003210] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.003400] env[62066]: DEBUG nova.virt.hardware [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.004352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b611b4e5-3a66-431c-84a4-94d0210a43bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.016171] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b977b782-3397-465b-946b-9d9df86a18d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.034832] env[62066]: ERROR nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Traceback (most recent call last): [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] yield resources [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.driver.spawn(context, instance, image_meta, [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] vm_ref = self.build_virtual_machine(instance, [ 609.034832] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] for vif in network_info: [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return self._sync_wrapper(fn, *args, **kwargs) [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.wait() [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self[:] = self._gt.wait() [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return self._exit_event.wait() [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 609.035199] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] current.throw(*self._exc) [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] result = function(*args, **kwargs) [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return func(*args, **kwargs) [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise e [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] nwinfo = self.network_api.allocate_for_instance( [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] created_port_ids = self._update_ports_for_instance( [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] with excutils.save_and_reraise_exception(): [ 609.035536] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.force_reraise() [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise self.value [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] updated_port = self._update_port( [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] _ensure_no_port_binding_failure(port) [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise exception.PortBindingFailed(port_id=port['id']) [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 609.035880] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] [ 609.035880] env[62066]: INFO nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Terminating instance [ 609.037437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.038156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.096s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.038729] env[62066]: ERROR nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Traceback (most recent call last): [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.driver.spawn(context, instance, image_meta, [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] vm_ref = self.build_virtual_machine(instance, [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.038729] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] for vif in network_info: [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self._sync_wrapper(fn, *args, **kwargs) [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.wait() [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self[:] = self._gt.wait() [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self._exit_event.wait() [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] result = hub.switch() [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.039071] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return self.greenlet.switch() [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] result = function(*args, **kwargs) [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] return func(*args, **kwargs) [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise e [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] nwinfo = self.network_api.allocate_for_instance( [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] created_port_ids = self._update_ports_for_instance( [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] with excutils.save_and_reraise_exception(): [ 609.039422] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] self.force_reraise() [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise self.value [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] updated_port = self._update_port( [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] _ensure_no_port_binding_failure(port) [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] raise exception.PortBindingFailed(port_id=port['id']) [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] nova.exception.PortBindingFailed: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. [ 609.039836] env[62066]: ERROR nova.compute.manager [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] [ 609.040171] env[62066]: DEBUG nova.compute.utils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 609.040505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquiring lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.040720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.720s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.042096] env[62066]: INFO nova.compute.claims [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.044529] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Build of instance 30f7baa5-0288-4ef3-9d1c-d7143868b3a4 was re-scheduled: Binding failed for port 1e296eab-a517-4d4f-b471-1a3042b5a5bd, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.044923] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.045145] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquiring lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.045290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Acquired lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.045446] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.110606] env[62066]: DEBUG nova.network.neutron [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.190335] env[62066]: DEBUG nova.network.neutron [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.565820] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.639220] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.694796] env[62066]: DEBUG oslo_concurrency.lockutils [req-4195d770-6146-4a26-bd8a-449a8c58b877 req-74230ad6-3c27-42c0-a36a-2c93c0c36df5 service nova] Releasing lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.695238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquired lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.695417] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.142342] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Releasing lock "refresh_cache-30f7baa5-0288-4ef3-9d1c-d7143868b3a4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.142342] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.142342] env[62066]: DEBUG nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.142342] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.159544] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.215048] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.288054] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.423955] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae225d93-2e4a-45c5-b84e-2ed703fdaa13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.431778] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e756f859-a53d-4ca2-966b-6baf0e56fa92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.462821] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130935d7-774b-4fb4-a7c5-7a660df07058 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.470201] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e7c103-851f-4ea8-bc97-51f8479ff662 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.483152] env[62066]: DEBUG nova.compute.provider_tree [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.637425] env[62066]: DEBUG nova.compute.manager [req-743ac17d-5f38-4dad-9e6c-0b5e9cf0f454 req-f0ca89ca-49c1-48d9-87cc-e22d18c7454f service nova] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Received event network-vif-deleted-534dfc23-9752-49e0-a34a-c21f75401d5e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.662342] env[62066]: DEBUG nova.network.neutron [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.790639] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Releasing lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.790905] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.791055] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.791383] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d69cba7-0e9a-4db7-a4bd-4521ad8862b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.800331] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c546dd3e-8741-47be-82a9-b3d06db75cd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.821753] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1 could not be found. [ 610.821968] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.822163] env[62066]: INFO nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 610.822399] env[62066]: DEBUG oslo.service.loopingcall [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.822604] env[62066]: DEBUG nova.compute.manager [-] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.822695] env[62066]: DEBUG nova.network.neutron [-] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.838847] env[62066]: DEBUG nova.network.neutron [-] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.986579] env[62066]: DEBUG nova.scheduler.client.report [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.165133] env[62066]: INFO nova.compute.manager [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] [instance: 30f7baa5-0288-4ef3-9d1c-d7143868b3a4] Took 1.02 seconds to deallocate network for instance. [ 611.341238] env[62066]: DEBUG nova.network.neutron [-] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.492559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.493360] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.495575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.165s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.497289] env[62066]: INFO nova.compute.claims [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.843642] env[62066]: INFO nova.compute.manager [-] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Took 1.02 seconds to deallocate network for instance. [ 611.845966] env[62066]: DEBUG nova.compute.claims [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.846156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.001614] env[62066]: DEBUG nova.compute.utils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 612.006111] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 612.195649] env[62066]: INFO nova.scheduler.client.report [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Deleted allocations for instance 30f7baa5-0288-4ef3-9d1c-d7143868b3a4 [ 612.507311] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.702884] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0ae2119-01d0-4648-a65f-9c9aa165a78d tempest-ServersAdminTestJSON-1636493553 tempest-ServersAdminTestJSON-1636493553-project-member] Lock "30f7baa5-0288-4ef3-9d1c-d7143868b3a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.871s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.868739] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5e948e-b551-4b82-8f94-030f7b80636d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.874701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3835a26e-34ed-4b69-8610-30afdf2e5655 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.904919] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7a81f6-20db-416d-80fb-5198778746d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.911992] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624fff9e-7337-4f34-8539-1e65a19cec38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.925791] env[62066]: DEBUG nova.compute.provider_tree [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.205471] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.428524] env[62066]: DEBUG nova.scheduler.client.report [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.521250] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.547067] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.547337] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.547504] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.547644] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.547786] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.548024] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.548673] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.548673] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.548673] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.548673] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.548934] env[62066]: DEBUG nova.virt.hardware [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.549665] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b2b36f-e1c7-421f-89ff-1ada1e06865c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.557562] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74665dcf-50ff-4048-b76d-51626996fb95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.572018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 613.576477] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating folder: Project (b8302dd17050453d9d8cfddc201e9757). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.577107] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3dc49df-0816-4c3d-a8a8-b5b3398e29a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.587189] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created folder: Project (b8302dd17050453d9d8cfddc201e9757) in parent group-v251573. [ 613.587391] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating folder: Instances. Parent ref: group-v251581. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.587620] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ab810db-7e6d-4c66-8282-a902548b6ec8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.597996] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created folder: Instances in parent group-v251581. [ 613.597996] env[62066]: DEBUG oslo.service.loopingcall [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.597996] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 613.597996] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-313faf44-ee29-406a-a106-cf4434d508de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.615771] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 613.615771] env[62066]: value = "task-1155674" [ 613.615771] env[62066]: _type = "Task" [ 613.615771] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.623408] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155674, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.728438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.933688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.934233] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 613.938256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.619s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.939742] env[62066]: INFO nova.compute.claims [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.126504] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155674, 'name': CreateVM_Task, 'duration_secs': 0.244547} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.126691] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 614.127130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.127296] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.127594] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 614.127831] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bbbf8c1-3ada-44cb-85e9-10cea1cc6b4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.132319] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 614.132319] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52adf2eb-090f-ca24-5c25-56bab76c03b4" [ 614.132319] env[62066]: _type = "Task" [ 614.132319] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.139795] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52adf2eb-090f-ca24-5c25-56bab76c03b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.151079] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 614.151079] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 614.439706] env[62066]: DEBUG nova.compute.utils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 614.441534] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.441762] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 614.488623] env[62066]: DEBUG nova.policy [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f453531ab9864e8ba056c5cf2dbfb603', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '120bc7d4eb004ee3982fad8283f6f885', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.649022] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52adf2eb-090f-ca24-5c25-56bab76c03b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.649975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.650244] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 614.650519] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.650679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.650841] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 614.651128] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2746dde-34b2-431e-aff9-6f792b94ece8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.656250] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 614.656408] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 614.656541] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 614.662268] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 614.664926] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 614.664926] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9c7e187-4507-4ce9-b50f-e628e7c62491 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.673743] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 614.673743] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52287025-a202-c6e1-f5dc-7a9c4c589b21" [ 614.673743] env[62066]: _type = "Task" [ 614.673743] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.681865] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52287025-a202-c6e1-f5dc-7a9c4c589b21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.743306] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Successfully created port: a56380e7-deee-408b-9b19-5661ee095742 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.945713] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: de135f73-53d2-4471-a0ba-9204af726690] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.164863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 615.165042] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 615.165042] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165204] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165321] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165480] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165602] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165743] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.165868] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 615.166012] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.189241] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52287025-a202-c6e1-f5dc-7a9c4c589b21, 'name': SearchDatastore_Task, 'duration_secs': 0.008196} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.192520] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-268afcef-5f1e-4f37-ac7b-194b460618f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.198150] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 615.198150] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5212d2a4-0868-0c3f-af5c-5dd3b7316927" [ 615.198150] env[62066]: _type = "Task" [ 615.198150] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.206333] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5212d2a4-0868-0c3f-af5c-5dd3b7316927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.397116] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d87c167-e8b3-4a59-b540-0a52f6fba21e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.405265] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fdeb7d-ec04-4a96-9f3e-8b9dff9f2ab9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.437516] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab70a309-b64e-44cf-bf01-baa24f756aa1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.445628] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b759d8f6-23b0-4360-a18e-bf77832d91b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.462980] env[62066]: DEBUG nova.compute.provider_tree [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.609836] env[62066]: DEBUG nova.compute.manager [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Received event network-changed-a56380e7-deee-408b-9b19-5661ee095742 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.610076] env[62066]: DEBUG nova.compute.manager [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Refreshing instance network info cache due to event network-changed-a56380e7-deee-408b-9b19-5661ee095742. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.610270] env[62066]: DEBUG oslo_concurrency.lockutils [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] Acquiring lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.610414] env[62066]: DEBUG oslo_concurrency.lockutils [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] Acquired lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.610570] env[62066]: DEBUG nova.network.neutron [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Refreshing network info cache for port a56380e7-deee-408b-9b19-5661ee095742 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 615.670009] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.709170] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5212d2a4-0868-0c3f-af5c-5dd3b7316927, 'name': SearchDatastore_Task, 'duration_secs': 0.008367} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.709467] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.709723] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b4cb3d6d-d8ec-4632-8020-fae39b18143e/b4cb3d6d-d8ec-4632-8020-fae39b18143e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 615.710035] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8791c2d9-d651-4be9-a3ae-445426bf997b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.716743] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 615.716743] env[62066]: value = "task-1155675" [ 615.716743] env[62066]: _type = "Task" [ 615.716743] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.724441] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.752913] env[62066]: ERROR nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 615.752913] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.752913] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.752913] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.752913] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.752913] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.752913] env[62066]: ERROR nova.compute.manager raise self.value [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.752913] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.752913] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.752913] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.753401] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.753401] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.753401] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 615.753401] env[62066]: ERROR nova.compute.manager [ 615.753401] env[62066]: Traceback (most recent call last): [ 615.753401] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.753401] env[62066]: listener.cb(fileno) [ 615.753401] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.753401] env[62066]: result = function(*args, **kwargs) [ 615.753401] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.753401] env[62066]: return func(*args, **kwargs) [ 615.753401] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.753401] env[62066]: raise e [ 615.753401] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.753401] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 615.753401] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.753401] env[62066]: created_port_ids = self._update_ports_for_instance( [ 615.753401] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.753401] env[62066]: with excutils.save_and_reraise_exception(): [ 615.753401] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.753401] env[62066]: self.force_reraise() [ 615.753401] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.753401] env[62066]: raise self.value [ 615.753401] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.753401] env[62066]: updated_port = self._update_port( [ 615.753401] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.753401] env[62066]: _ensure_no_port_binding_failure(port) [ 615.753401] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.753401] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.754081] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 615.754081] env[62066]: Removing descriptor: 19 [ 615.968033] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 615.970640] env[62066]: DEBUG nova.scheduler.client.report [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.015601] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.015902] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.016071] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.016580] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.016580] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.016744] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.016818] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.017176] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.017176] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.017328] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.017491] env[62066]: DEBUG nova.virt.hardware [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.019284] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8958210-a445-4ec8-a6f6-179fd2b80e5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.028415] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69221a7c-825a-4f76-933b-8162fb341dcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.046105] env[62066]: ERROR nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Traceback (most recent call last): [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] yield resources [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.driver.spawn(context, instance, image_meta, [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] vm_ref = self.build_virtual_machine(instance, [ 616.046105] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] for vif in network_info: [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return self._sync_wrapper(fn, *args, **kwargs) [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.wait() [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self[:] = self._gt.wait() [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return self._exit_event.wait() [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.046866] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] current.throw(*self._exc) [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] result = function(*args, **kwargs) [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return func(*args, **kwargs) [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise e [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] nwinfo = self.network_api.allocate_for_instance( [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] created_port_ids = self._update_ports_for_instance( [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] with excutils.save_and_reraise_exception(): [ 616.047377] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.force_reraise() [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise self.value [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] updated_port = self._update_port( [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] _ensure_no_port_binding_failure(port) [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise exception.PortBindingFailed(port_id=port['id']) [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 616.047951] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] [ 616.047951] env[62066]: INFO nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Terminating instance [ 616.050031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.133224] env[62066]: DEBUG nova.network.neutron [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.228233] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461859} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.228503] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b4cb3d6d-d8ec-4632-8020-fae39b18143e/b4cb3d6d-d8ec-4632-8020-fae39b18143e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 616.228707] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 616.228946] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fe23daf-b65e-4ae2-b592-9509e98bca71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.232959] env[62066]: DEBUG nova.network.neutron [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.239541] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 616.239541] env[62066]: value = "task-1155676" [ 616.239541] env[62066]: _type = "Task" [ 616.239541] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.250569] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.476090] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.476594] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.482501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.153s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.483909] env[62066]: INFO nova.compute.claims [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.735488] env[62066]: DEBUG oslo_concurrency.lockutils [req-467ee79c-9fac-488c-b5c9-e36120792c31 req-5bc717c7-7511-4122-82e1-79e7b28a87a8 service nova] Releasing lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.735910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.736120] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.754147] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059314} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.754407] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 616.755159] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea38b521-a33e-4b97-bfc6-5893ec43b7f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.775848] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] b4cb3d6d-d8ec-4632-8020-fae39b18143e/b4cb3d6d-d8ec-4632-8020-fae39b18143e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 616.776384] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93a6a78c-1381-44d0-a9c1-2a10f9ddcd41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.796709] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 616.796709] env[62066]: value = "task-1155677" [ 616.796709] env[62066]: _type = "Task" [ 616.796709] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.804353] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155677, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.988513] env[62066]: DEBUG nova.compute.utils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 616.996015] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.996217] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 617.056990] env[62066]: DEBUG nova.policy [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b786921efb3540828b1079769c40059d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd06323d3542b4e5c9d61f34173da6895', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.306148] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155677, 'name': ReconfigVM_Task, 'duration_secs': 0.500589} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.306459] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Reconfigured VM instance instance-00000016 to attach disk [datastore2] b4cb3d6d-d8ec-4632-8020-fae39b18143e/b4cb3d6d-d8ec-4632-8020-fae39b18143e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 617.307198] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b66b80ed-ed13-4260-b2ea-960866e9d354 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.312396] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 617.312396] env[62066]: value = "task-1155678" [ 617.312396] env[62066]: _type = "Task" [ 617.312396] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.320567] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155678, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.399274] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.497277] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.529943] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.542322] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Successfully created port: 3708b533-8f03-4217-a772-aecc26bd1bcc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.678065] env[62066]: DEBUG nova.compute.manager [req-18a3d5d0-a929-4c6a-92e9-62998ece58ba req-51a193b3-ba2b-4521-a96a-eef54e965d66 service nova] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Received event network-vif-deleted-a56380e7-deee-408b-9b19-5661ee095742 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.823450] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155678, 'name': Rename_Task, 'duration_secs': 0.136191} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.823450] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 617.823450] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9efd64b6-821f-4214-b043-dc4f61ea4a3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.829809] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 617.829809] env[62066]: value = "task-1155679" [ 617.829809] env[62066]: _type = "Task" [ 617.829809] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.846072] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.945225] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359113c5-6800-4aa5-81e8-d6677e621910 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.958192] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5289ecfc-9fa0-420d-a78d-00935a042bf9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.986849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc5ddb3-4127-491f-8e7b-102f236df3f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.994723] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e98aafe-92f2-4d0e-bc7e-9afb56079c59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.012367] env[62066]: DEBUG nova.compute.provider_tree [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.033281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.033281] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.033281] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.033281] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82dc1467-48c5-47d8-95bc-8d9f9eb30096 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.042959] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a8dde8-8008-40c8-b1ee-4ee6c014772c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.067472] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b5610b23-828b-4ece-b18a-4bfe27a60bbe could not be found. [ 618.067644] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.067896] env[62066]: INFO nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.068261] env[62066]: DEBUG oslo.service.loopingcall [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.069297] env[62066]: DEBUG nova.compute.manager [-] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.069457] env[62066]: DEBUG nova.network.neutron [-] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.090369] env[62066]: DEBUG nova.network.neutron [-] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.346320] env[62066]: DEBUG oslo_vmware.api [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155679, 'name': PowerOnVM_Task, 'duration_secs': 0.502354} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.346770] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 618.347120] env[62066]: INFO nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Took 4.83 seconds to spawn the instance on the hypervisor. [ 618.347435] env[62066]: DEBUG nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 618.348455] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e009e09a-a3ab-4c5b-9e0e-0a33cdf9e00e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.495376] env[62066]: ERROR nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 618.495376] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.495376] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.495376] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.495376] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.495376] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.495376] env[62066]: ERROR nova.compute.manager raise self.value [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.495376] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.495376] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.495376] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.496071] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.496071] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.496071] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 618.496071] env[62066]: ERROR nova.compute.manager [ 618.496071] env[62066]: Traceback (most recent call last): [ 618.496071] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.496071] env[62066]: listener.cb(fileno) [ 618.496071] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.496071] env[62066]: result = function(*args, **kwargs) [ 618.496071] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.496071] env[62066]: return func(*args, **kwargs) [ 618.496071] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.496071] env[62066]: raise e [ 618.496071] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.496071] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 618.496071] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.496071] env[62066]: created_port_ids = self._update_ports_for_instance( [ 618.496071] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.496071] env[62066]: with excutils.save_and_reraise_exception(): [ 618.496071] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.496071] env[62066]: self.force_reraise() [ 618.496071] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.496071] env[62066]: raise self.value [ 618.496071] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.496071] env[62066]: updated_port = self._update_port( [ 618.496071] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.496071] env[62066]: _ensure_no_port_binding_failure(port) [ 618.496071] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.496071] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.496735] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 618.496735] env[62066]: Removing descriptor: 19 [ 618.515094] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.517691] env[62066]: DEBUG nova.scheduler.client.report [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.543506] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.546855] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.546855] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.546855] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.546855] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.546855] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.547209] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.547209] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.547209] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.547209] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.547209] env[62066]: DEBUG nova.virt.hardware [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.547341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dbc45d-50f9-4723-afa0-e3f256a89f2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.554253] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9009ae-e0e4-46ec-b2fc-549e3cf23856 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.567892] env[62066]: ERROR nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Traceback (most recent call last): [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] yield resources [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.driver.spawn(context, instance, image_meta, [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] vm_ref = self.build_virtual_machine(instance, [ 618.567892] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] for vif in network_info: [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return self._sync_wrapper(fn, *args, **kwargs) [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.wait() [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self[:] = self._gt.wait() [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return self._exit_event.wait() [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.568470] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] current.throw(*self._exc) [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] result = function(*args, **kwargs) [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return func(*args, **kwargs) [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise e [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] nwinfo = self.network_api.allocate_for_instance( [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] created_port_ids = self._update_ports_for_instance( [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] with excutils.save_and_reraise_exception(): [ 618.569026] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.force_reraise() [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise self.value [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] updated_port = self._update_port( [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] _ensure_no_port_binding_failure(port) [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise exception.PortBindingFailed(port_id=port['id']) [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 618.569644] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] [ 618.570852] env[62066]: INFO nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Terminating instance [ 618.572685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.573066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquired lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.573543] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.592580] env[62066]: DEBUG nova.network.neutron [-] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.866818] env[62066]: INFO nova.compute.manager [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Took 27.56 seconds to build instance. [ 619.022511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.023357] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.026116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.558s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.091467] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.094891] env[62066]: INFO nova.compute.manager [-] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Took 1.03 seconds to deallocate network for instance. [ 619.097059] env[62066]: DEBUG nova.compute.claims [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.097274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.165309] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.369470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ca255e72-b848-4bc0-acc7-42c65f864b36 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.566s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.530935] env[62066]: DEBUG nova.compute.utils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 619.535477] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 619.667454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Releasing lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.667881] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.668179] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.668469] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9477a52-790a-4dec-a0a4-f82fe0ec2f6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.679141] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a71325-d7ae-4467-9acc-27fe0bb61fd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.703571] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5fa586bb-bf1c-4126-a2a7-70e488307794 could not be found. [ 619.703792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.703968] env[62066]: INFO nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Took 0.04 seconds to destroy the instance on the hypervisor. [ 619.704229] env[62066]: DEBUG oslo.service.loopingcall [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.707656] env[62066]: DEBUG nova.compute.manager [-] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.707757] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.710018] env[62066]: DEBUG nova.compute.manager [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Received event network-changed-3708b533-8f03-4217-a772-aecc26bd1bcc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.710288] env[62066]: DEBUG nova.compute.manager [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Refreshing instance network info cache due to event network-changed-3708b533-8f03-4217-a772-aecc26bd1bcc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 619.710422] env[62066]: DEBUG oslo_concurrency.lockutils [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] Acquiring lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.710575] env[62066]: DEBUG oslo_concurrency.lockutils [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] Acquired lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.710731] env[62066]: DEBUG nova.network.neutron [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Refreshing network info cache for port 3708b533-8f03-4217-a772-aecc26bd1bcc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 619.730257] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.875468] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.958580] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b3e2b5-f5e5-4936-9a8f-068b14db8d6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.966944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2861bb-6f93-4800-bd1a-7fbd1fd07c6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.999472] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5438d0cd-bcc8-474a-8a59-98c7308d7436 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.007560] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6f5325-bc77-44c0-b8b2-2ec183d90e32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.023365] env[62066]: DEBUG nova.compute.provider_tree [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.038325] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.233120] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.240237] env[62066]: DEBUG nova.network.neutron [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.364545] env[62066]: DEBUG nova.network.neutron [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.399228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.527337] env[62066]: DEBUG nova.scheduler.client.report [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.735587] env[62066]: INFO nova.compute.manager [-] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Took 1.03 seconds to deallocate network for instance. [ 620.738518] env[62066]: DEBUG nova.compute.claims [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.738697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.867467] env[62066]: DEBUG oslo_concurrency.lockutils [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] Releasing lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.867769] env[62066]: DEBUG nova.compute.manager [req-59c90c96-501e-4e6e-97a6-e061c2c922ed req-3fc659ea-ef8f-4cfc-95a3-f1ca8ee547cd service nova] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Received event network-vif-deleted-3708b533-8f03-4217-a772-aecc26bd1bcc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.033498] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.007s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.034587] env[62066]: ERROR nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] Traceback (most recent call last): [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.driver.spawn(context, instance, image_meta, [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] vm_ref = self.build_virtual_machine(instance, [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.034587] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] for vif in network_info: [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self._sync_wrapper(fn, *args, **kwargs) [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.wait() [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self[:] = self._gt.wait() [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self._exit_event.wait() [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] result = hub.switch() [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.035245] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return self.greenlet.switch() [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] result = function(*args, **kwargs) [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] return func(*args, **kwargs) [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise e [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] nwinfo = self.network_api.allocate_for_instance( [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] created_port_ids = self._update_ports_for_instance( [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] with excutils.save_and_reraise_exception(): [ 621.035719] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] self.force_reraise() [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise self.value [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] updated_port = self._update_port( [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] _ensure_no_port_binding_failure(port) [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] raise exception.PortBindingFailed(port_id=port['id']) [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] nova.exception.PortBindingFailed: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. [ 621.036027] env[62066]: ERROR nova.compute.manager [instance: de135f73-53d2-4471-a0ba-9204af726690] [ 621.036286] env[62066]: DEBUG nova.compute.utils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 621.037521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.209s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.039692] env[62066]: INFO nova.compute.claims [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.043944] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Build of instance de135f73-53d2-4471-a0ba-9204af726690 was re-scheduled: Binding failed for port ec358e5f-1fb3-406a-98ca-75864e8177e0, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.044580] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.044864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquiring lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.045077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Acquired lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.045315] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.050873] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.082935] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.085017] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.085564] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.085564] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.085564] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.085564] env[62066]: DEBUG nova.virt.hardware [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.085867] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23594e50-0276-4547-abd5-0263de3cc4cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.094511] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56997ce-1df4-496d-9599-3d50b21c7817 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.110121] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 621.116182] env[62066]: DEBUG oslo.service.loopingcall [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.116401] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 621.116691] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8464ee6-4410-43e9-bcb3-173c7cb72ef9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.133797] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 621.133797] env[62066]: value = "task-1155680" [ 621.133797] env[62066]: _type = "Task" [ 621.133797] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.141840] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155680, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.571343] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.619586] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.619828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.631183] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.645176] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155680, 'name': CreateVM_Task, 'duration_secs': 0.26517} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.645353] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 621.645778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.645928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.646525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 621.646779] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee354921-f32f-455c-969b-6b64a7230424 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.651379] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 621.651379] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bfc3a1-d956-2190-3543-fd7decfbe141" [ 621.651379] env[62066]: _type = "Task" [ 621.651379] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.662995] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bfc3a1-d956-2190-3543-fd7decfbe141, 'name': SearchDatastore_Task, 'duration_secs': 0.008675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.663289] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.663515] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.663739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.663881] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.664064] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 621.664564] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d59a8dc3-2de9-4a7c-bc3a-4bf02b151950 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.672095] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 621.672269] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 621.672992] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f15406e1-2657-4f9f-b59f-ad42e40f663e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.678320] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 621.678320] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521cffc6-bdee-1799-6c98-ce56d9b6b6d3" [ 621.678320] env[62066]: _type = "Task" [ 621.678320] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.687172] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521cffc6-bdee-1799-6c98-ce56d9b6b6d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.135922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Releasing lock "refresh_cache-de135f73-53d2-4471-a0ba-9204af726690" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.136197] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.136335] env[62066]: DEBUG nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.136495] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.153192] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.189605] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521cffc6-bdee-1799-6c98-ce56d9b6b6d3, 'name': SearchDatastore_Task, 'duration_secs': 0.007605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.195030] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-583a8ebd-57e7-4a49-8dcb-8d022e756930 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.197711] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 622.197711] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b5d20b-6c6d-82e4-00fe-242813eefab5" [ 622.197711] env[62066]: _type = "Task" [ 622.197711] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.206538] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b5d20b-6c6d-82e4-00fe-242813eefab5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.428639] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebb3fec-8d8b-4792-b808-72cdf918d4d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.437352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6004e842-7334-4e21-b896-72dec7b1a910 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.467449] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d789e2d6-b4b8-4152-8625-4d809baeae30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.474696] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f375a1c-7e32-47cc-9ddc-3a1079a41037 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.487842] env[62066]: DEBUG nova.compute.provider_tree [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.660018] env[62066]: DEBUG nova.network.neutron [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.709383] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b5d20b-6c6d-82e4-00fe-242813eefab5, 'name': SearchDatastore_Task, 'duration_secs': 0.00854} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.709645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.709913] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 622.710155] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71b424b9-5524-4114-abf9-4ea8b90df171 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.716570] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 622.716570] env[62066]: value = "task-1155681" [ 622.716570] env[62066]: _type = "Task" [ 622.716570] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.724489] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.994351] env[62066]: DEBUG nova.scheduler.client.report [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.161847] env[62066]: INFO nova.compute.manager [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] [instance: de135f73-53d2-4471-a0ba-9204af726690] Took 1.03 seconds to deallocate network for instance. [ 623.227071] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155681, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447388} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.227344] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 623.227556] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 623.227799] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f1b6be9-683e-4b80-bded-7b432995dea3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.233788] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 623.233788] env[62066]: value = "task-1155682" [ 623.233788] env[62066]: _type = "Task" [ 623.233788] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.241912] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155682, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.499665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.500160] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.503050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.064s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.743246] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155682, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059232} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.743528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 623.744419] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0658f72-c1df-46a1-89aa-dcc4e75c8bba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.764537] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 623.765124] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed76bf8d-20d6-4034-a0ca-061f2e8c7606 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.783847] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 623.783847] env[62066]: value = "task-1155683" [ 623.783847] env[62066]: _type = "Task" [ 623.783847] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.794260] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155683, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.007628] env[62066]: DEBUG nova.compute.utils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 624.009364] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 624.186600] env[62066]: INFO nova.scheduler.client.report [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Deleted allocations for instance de135f73-53d2-4471-a0ba-9204af726690 [ 624.300022] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155683, 'name': ReconfigVM_Task, 'duration_secs': 0.261241} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.300022] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Reconfigured VM instance instance-00000019 to attach disk [datastore2] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 624.301118] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f9bc006-91cf-427a-b9db-42bfac758b5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.308553] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 624.308553] env[62066]: value = "task-1155684" [ 624.308553] env[62066]: _type = "Task" [ 624.308553] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.323602] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155684, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.371654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12176495-b653-4f1e-9df7-d4190833f79d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.378856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd96e244-4165-4c0f-9856-0a61ec94a4f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.409812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0502b517-0ede-4b4d-9860-708e35c174a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.417787] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0992ca33-b74f-4f79-b394-228fdd08989f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.431487] env[62066]: DEBUG nova.compute.provider_tree [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.514242] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.696628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-af1e3d96-0c28-4a63-8f9c-86f8b4e4aafd tempest-TenantUsagesTestJSON-2032850452 tempest-TenantUsagesTestJSON-2032850452-project-member] Lock "de135f73-53d2-4471-a0ba-9204af726690" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.289s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.821034] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155684, 'name': Rename_Task, 'duration_secs': 0.128215} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.821191] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 624.821432] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5085ab22-d557-4f53-8ff0-431693468d97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.828065] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 624.828065] env[62066]: value = "task-1155685" [ 624.828065] env[62066]: _type = "Task" [ 624.828065] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.835496] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155685, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.935203] env[62066]: DEBUG nova.scheduler.client.report [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.200513] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.337556] env[62066]: DEBUG oslo_vmware.api [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155685, 'name': PowerOnVM_Task, 'duration_secs': 0.397128} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.337823] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.338706] env[62066]: INFO nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Took 4.29 seconds to spawn the instance on the hypervisor. [ 625.338880] env[62066]: DEBUG nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 625.339861] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed6ac42-85d2-4ccd-b685-cf382cab97b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.440574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.441264] env[62066]: ERROR nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Traceback (most recent call last): [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.driver.spawn(context, instance, image_meta, [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] vm_ref = self.build_virtual_machine(instance, [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.441264] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] for vif in network_info: [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self._sync_wrapper(fn, *args, **kwargs) [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.wait() [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self[:] = self._gt.wait() [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self._exit_event.wait() [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] result = hub.switch() [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.441558] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return self.greenlet.switch() [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] result = function(*args, **kwargs) [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] return func(*args, **kwargs) [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise e [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] nwinfo = self.network_api.allocate_for_instance( [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] created_port_ids = self._update_ports_for_instance( [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] with excutils.save_and_reraise_exception(): [ 625.441855] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] self.force_reraise() [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise self.value [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] updated_port = self._update_port( [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] _ensure_no_port_binding_failure(port) [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] raise exception.PortBindingFailed(port_id=port['id']) [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] nova.exception.PortBindingFailed: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. [ 625.442434] env[62066]: ERROR nova.compute.manager [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] [ 625.443779] env[62066]: DEBUG nova.compute.utils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 625.443779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.406s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.444743] env[62066]: INFO nova.compute.claims [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.449083] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Build of instance c8bad79f-9bf7-4ef2-b681-ca8901e878fd was re-scheduled: Binding failed for port 309971aa-38f4-4b36-b889-eb6ee86b8200, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.449083] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.449083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquiring lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.449261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Acquired lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.449298] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.529502] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.553668] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.553903] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.554073] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.554258] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.554403] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.554546] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.554750] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.554903] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.555173] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.555371] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.555565] env[62066]: DEBUG nova.virt.hardware [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.556446] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c911cb46-aa06-448e-ae0a-f651be6cae2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.564532] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8cf9f5-105f-4194-9f5c-22c7b7b85896 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.577664] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 625.583209] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Creating folder: Project (2d08cd87db7e4ab4b1d3677989470afc). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 625.583493] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c19a2ba-c075-485f-afb3-1e01dc3faeee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.592814] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Created folder: Project (2d08cd87db7e4ab4b1d3677989470afc) in parent group-v251573. [ 625.593008] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Creating folder: Instances. Parent ref: group-v251585. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 625.593251] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e5c9f92-ef4a-4d7f-9346-22d2d2a9f819 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.601057] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Created folder: Instances in parent group-v251585. [ 625.601280] env[62066]: DEBUG oslo.service.loopingcall [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.601461] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 625.601646] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66eb0c7d-3c0f-4af1-b954-81f741701448 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.617713] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 625.617713] env[62066]: value = "task-1155688" [ 625.617713] env[62066]: _type = "Task" [ 625.617713] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.625235] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155688, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.722947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.859868] env[62066]: INFO nova.compute.manager [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Took 28.56 seconds to build instance. [ 625.969222] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.057776] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.127993] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155688, 'name': CreateVM_Task, 'duration_secs': 0.275504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.128212] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 626.128633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.128793] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.129125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 626.129383] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbe5afbe-87f5-4712-be18-13d5838b79d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.133597] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 626.133597] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e59a31-de74-e064-69b7-61e668c81713" [ 626.133597] env[62066]: _type = "Task" [ 626.133597] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.142730] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e59a31-de74-e064-69b7-61e668c81713, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.362197] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b95388e7-b381-45bc-b2c8-1b59efcc126d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.575s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.561430] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Releasing lock "refresh_cache-c8bad79f-9bf7-4ef2-b681-ca8901e878fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.561670] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.561852] env[62066]: DEBUG nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.562031] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.578251] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.646507] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e59a31-de74-e064-69b7-61e668c81713, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.647126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.647126] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 626.647324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.647473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.647651] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 626.647897] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3152a462-9fd0-4be9-9435-6e4ee39c52bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.661208] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.661418] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 626.664242] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acd21d09-1250-4633-9a78-e2a7a354faa0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.669563] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 626.669563] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a988c8-9dcf-68ed-7323-343eed7a2c66" [ 626.669563] env[62066]: _type = "Task" [ 626.669563] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.676987] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a988c8-9dcf-68ed-7323-343eed7a2c66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.731500] env[62066]: INFO nova.compute.manager [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Rebuilding instance [ 626.771612] env[62066]: DEBUG nova.compute.manager [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 626.772036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8a008e-4314-4ecc-9a1a-23a71b7885f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.812675] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca87b7b-268e-4d7d-b952-5e77f69ba5f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.820761] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3658f37d-4bdc-4837-a31a-0e2b05a9a699 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.852550] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be56260-058d-47ca-8c70-b511bbc70fd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.859513] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190f4d68-b8bd-4b56-be07-553ee3bfcc7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.865147] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.875304] env[62066]: DEBUG nova.compute.provider_tree [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.083661] env[62066]: DEBUG nova.network.neutron [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.179231] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a988c8-9dcf-68ed-7323-343eed7a2c66, 'name': SearchDatastore_Task, 'duration_secs': 0.035622} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.179983] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e60448f-1537-42f2-9124-9c74974d952d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.184853] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 627.184853] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52595c49-5ce2-d767-55be-907115e4a78f" [ 627.184853] env[62066]: _type = "Task" [ 627.184853] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.192546] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52595c49-5ce2-d767-55be-907115e4a78f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.285781] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 627.286108] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-350de602-0b9b-470d-b4ea-8eb703d42af1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.293523] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 627.293523] env[62066]: value = "task-1155689" [ 627.293523] env[62066]: _type = "Task" [ 627.293523] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.302047] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.380172] env[62066]: DEBUG nova.scheduler.client.report [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.398983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.585951] env[62066]: INFO nova.compute.manager [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] [instance: c8bad79f-9bf7-4ef2-b681-ca8901e878fd] Took 1.02 seconds to deallocate network for instance. [ 627.695473] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52595c49-5ce2-d767-55be-907115e4a78f, 'name': SearchDatastore_Task, 'duration_secs': 0.021823} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.695725] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.695983] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 627.696311] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82e62250-b69b-44ff-9e8c-c7f2cac65134 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.702922] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 627.702922] env[62066]: value = "task-1155690" [ 627.702922] env[62066]: _type = "Task" [ 627.702922] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.711237] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155690, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.803553] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155689, 'name': PowerOffVM_Task, 'duration_secs': 0.222742} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.803879] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 627.804150] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.805206] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032b2532-40c1-484b-8242-0742b063620b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.811728] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 627.811960] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3821da23-9a80-43f3-9d5d-b8d7c9f762a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.841974] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 627.842241] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 627.842437] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleting the datastore file [datastore2] 2dc8d357-2864-410a-93ef-21ea4e9b530c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 627.842750] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-757680f4-6be1-4060-aa08-70fcc8afd972 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.849472] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 627.849472] env[62066]: value = "task-1155692" [ 627.849472] env[62066]: _type = "Task" [ 627.849472] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.861647] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155692, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.887136] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.888427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.042s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.213624] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155690, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.213882] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 628.214114] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 628.214389] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c993f7c-b4f6-4b74-84d3-c5030f41d852 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.221397] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 628.221397] env[62066]: value = "task-1155693" [ 628.221397] env[62066]: _type = "Task" [ 628.221397] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.229279] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155693, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.359659] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155692, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.364384} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.359854] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 628.360049] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 628.360299] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.392283] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "5878f1a5-40c6-4e93-b89c-2f0d2c1827d8" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.392546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "5878f1a5-40c6-4e93-b89c-2f0d2c1827d8" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.611086] env[62066]: INFO nova.scheduler.client.report [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Deleted allocations for instance c8bad79f-9bf7-4ef2-b681-ca8901e878fd [ 628.733489] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155693, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07219} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.733735] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.734551] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5763de3-9a1d-4bde-95c8-b5dcf98bfe19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.738209] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e959ba-5212-4b57-8918-aa0707bff57b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.751435] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa2f7c7-0405-4183-90dd-ae76f69c998e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.763565] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.764100] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a508a2af-68b7-4216-b36b-b3edf9bbd5cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.809137] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa38034-85c5-4a40-8885-eadf42a006f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.812110] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 628.812110] env[62066]: value = "task-1155694" [ 628.812110] env[62066]: _type = "Task" [ 628.812110] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.819228] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60f121b-d998-435d-b9ef-b77f70abcf7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.826450] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155694, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.837151] env[62066]: DEBUG nova.compute.provider_tree [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.898161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "5878f1a5-40c6-4e93-b89c-2f0d2c1827d8" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.898761] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.121635] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2943152-ac11-46e2-8fef-384cc941abc3 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601 tempest-FloatingIPsAssociationNegativeTestJSON-1891908601-project-member] Lock "c8bad79f-9bf7-4ef2-b681-ca8901e878fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.701s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.323500] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155694, 'name': ReconfigVM_Task, 'duration_secs': 0.358294} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.323500] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfigured VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.324014] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dcd49c9b-4637-47ac-bc9e-428113813a11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.330359] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 629.330359] env[62066]: value = "task-1155695" [ 629.330359] env[62066]: _type = "Task" [ 629.330359] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.337991] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155695, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.340885] env[62066]: DEBUG nova.scheduler.client.report [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.396610] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.396891] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.397019] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.397217] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.397375] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.397524] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.397729] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.397886] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.398063] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.398238] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.398404] env[62066]: DEBUG nova.virt.hardware [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.399264] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7022c6a5-7e9f-478a-91ad-09cc88ac1b14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.404101] env[62066]: DEBUG nova.compute.utils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 629.407358] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.407528] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.410466] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1397ea-901a-4da4-842b-dd197e8f45d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.426199] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.431788] env[62066]: DEBUG oslo.service.loopingcall [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.432336] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 629.432575] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f18c6429-65fa-4758-b58b-64931117748c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.449216] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.449216] env[62066]: value = "task-1155696" [ 629.449216] env[62066]: _type = "Task" [ 629.449216] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.459965] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155696, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.490810] env[62066]: DEBUG nova.policy [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26cf67005dba475a9d4bfe8b699f7620', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '923ba26b64974a419b3ef484aa568749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.623404] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.845332] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155695, 'name': Rename_Task, 'duration_secs': 0.132965} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.846071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.846693] env[62066]: ERROR nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Traceback (most recent call last): [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.driver.spawn(context, instance, image_meta, [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] vm_ref = self.build_virtual_machine(instance, [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.846693] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] for vif in network_info: [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return self._sync_wrapper(fn, *args, **kwargs) [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.wait() [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self[:] = self._gt.wait() [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return self._exit_event.wait() [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] current.throw(*self._exc) [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.846989] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] result = function(*args, **kwargs) [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] return func(*args, **kwargs) [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise e [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] nwinfo = self.network_api.allocate_for_instance( [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] created_port_ids = self._update_ports_for_instance( [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] with excutils.save_and_reraise_exception(): [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] self.force_reraise() [ 629.847292] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise self.value [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] updated_port = self._update_port( [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] _ensure_no_port_binding_failure(port) [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] raise exception.PortBindingFailed(port_id=port['id']) [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] nova.exception.PortBindingFailed: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. [ 629.847586] env[62066]: ERROR nova.compute.manager [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] [ 629.847586] env[62066]: DEBUG nova.compute.utils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 629.848603] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.849090] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Build of instance 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1 was re-scheduled: Binding failed for port 534dfc23-9752-49e0-a34a-c21f75401d5e, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.849517] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.849747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquiring lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.849903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Acquired lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.850061] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.851010] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.123s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.852506] env[62066]: INFO nova.compute.claims [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.854716] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf6afc9f-85cd-416c-93ce-c07ae92b080a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.860880] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 629.860880] env[62066]: value = "task-1155697" [ 629.860880] env[62066]: _type = "Task" [ 629.860880] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.869721] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.910422] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.948732] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Successfully created port: d0da24d6-830a-484a-af48-e42459fd3b21 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.962325] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155696, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.151689] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.371626] env[62066]: DEBUG oslo_vmware.api [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155697, 'name': PowerOnVM_Task, 'duration_secs': 0.399934} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.372546] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.372877] env[62066]: INFO nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Took 4.84 seconds to spawn the instance on the hypervisor. [ 630.373192] env[62066]: DEBUG nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.374255] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ae2d44-bde6-4c15-990e-343dae258c79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.385249] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.460897] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155696, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.703684] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.896200] env[62066]: INFO nova.compute.manager [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Took 29.09 seconds to build instance. [ 630.920239] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.954403] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.954667] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.954905] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.955009] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.955186] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.955333] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.955535] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.955688] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.955850] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.956017] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.956605] env[62066]: DEBUG nova.virt.hardware [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.957935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0add5f12-51ed-48be-97cb-2e637327fb59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.972536] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155696, 'name': CreateVM_Task, 'duration_secs': 1.275341} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.975049] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 630.976046] env[62066]: DEBUG oslo_vmware.service [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6176d3d2-cde0-41e1-9e2d-a1c22da8f288 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.980430] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729a124a-c98a-450b-86be-ef4e38df4951 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.992923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.993121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.993487] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 630.993724] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17d54b83-8699-40d1-9fe0-75ca0c50b908 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.009707] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 631.009707] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527a88eb-e62e-51d2-a8c0-3184ff8fc7af" [ 631.009707] env[62066]: _type = "Task" [ 631.009707] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.019709] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a88eb-e62e-51d2-a8c0-3184ff8fc7af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.210690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Releasing lock "refresh_cache-80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.211767] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.212737] env[62066]: DEBUG nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.215026] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.240023] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.296133] env[62066]: DEBUG nova.compute.manager [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Received event network-changed-d0da24d6-830a-484a-af48-e42459fd3b21 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.297105] env[62066]: DEBUG nova.compute.manager [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Refreshing instance network info cache due to event network-changed-d0da24d6-830a-484a-af48-e42459fd3b21. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.297105] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] Acquiring lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.297105] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] Acquired lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.297105] env[62066]: DEBUG nova.network.neutron [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Refreshing network info cache for port d0da24d6-830a-484a-af48-e42459fd3b21 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 631.385916] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e021e6c6-3597-42e8-abb3-ce8def6e30e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.394237] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569b1b59-b2fe-4bc5-8887-d3dc0d22d7ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.399578] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80b84e7c-c2ae-4dfa-acbc-ebb31c9583a9 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.323s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.431939] env[62066]: ERROR nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 631.431939] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.431939] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.431939] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.431939] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.431939] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.431939] env[62066]: ERROR nova.compute.manager raise self.value [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.431939] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.431939] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.431939] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.432499] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.432499] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.432499] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 631.432499] env[62066]: ERROR nova.compute.manager [ 631.432499] env[62066]: Traceback (most recent call last): [ 631.432499] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.432499] env[62066]: listener.cb(fileno) [ 631.432499] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.432499] env[62066]: result = function(*args, **kwargs) [ 631.432499] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.432499] env[62066]: return func(*args, **kwargs) [ 631.432499] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.432499] env[62066]: raise e [ 631.432499] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.432499] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 631.432499] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.432499] env[62066]: created_port_ids = self._update_ports_for_instance( [ 631.432499] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.432499] env[62066]: with excutils.save_and_reraise_exception(): [ 631.432499] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.432499] env[62066]: self.force_reraise() [ 631.432499] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.432499] env[62066]: raise self.value [ 631.432499] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.432499] env[62066]: updated_port = self._update_port( [ 631.432499] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.432499] env[62066]: _ensure_no_port_binding_failure(port) [ 631.432499] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.432499] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.433203] env[62066]: nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 631.433203] env[62066]: Removing descriptor: 19 [ 631.433203] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a698bb-73c0-419d-bb69-8f2c3dc1a054 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.435839] env[62066]: ERROR nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Traceback (most recent call last): [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] yield resources [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.driver.spawn(context, instance, image_meta, [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] vm_ref = self.build_virtual_machine(instance, [ 631.435839] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] for vif in network_info: [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self._sync_wrapper(fn, *args, **kwargs) [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.wait() [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self[:] = self._gt.wait() [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self._exit_event.wait() [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.436196] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] result = hub.switch() [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self.greenlet.switch() [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] result = function(*args, **kwargs) [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return func(*args, **kwargs) [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise e [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] nwinfo = self.network_api.allocate_for_instance( [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] created_port_ids = self._update_ports_for_instance( [ 631.436515] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] with excutils.save_and_reraise_exception(): [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.force_reraise() [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise self.value [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] updated_port = self._update_port( [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] _ensure_no_port_binding_failure(port) [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise exception.PortBindingFailed(port_id=port['id']) [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 631.436834] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] [ 631.437171] env[62066]: INFO nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Terminating instance [ 631.438658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.448431] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f7019b-dfa0-44ff-a0c0-9f6fb1d08afc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.464212] env[62066]: DEBUG nova.compute.provider_tree [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.525013] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.525383] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 631.525601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.526101] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.526359] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 631.527069] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d1d4b43-0e02-4c4b-8f83-e255c5955cab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.546756] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 631.546954] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 631.547841] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fc1323-b428-4d21-bdc7-eed3f9e5df33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.554322] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ac9970a-caa8-4161-9be8-7eb2228ba386 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.559589] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 631.559589] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5276adcc-551f-2c52-4258-f2f6d7ad1fe2" [ 631.559589] env[62066]: _type = "Task" [ 631.559589] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.572885] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5276adcc-551f-2c52-4258-f2f6d7ad1fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.744572] env[62066]: DEBUG nova.network.neutron [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.816871] env[62066]: DEBUG nova.network.neutron [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.893630] env[62066]: DEBUG nova.network.neutron [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.905230] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.966902] env[62066]: DEBUG nova.scheduler.client.report [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.007317] env[62066]: INFO nova.compute.manager [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Rebuilding instance [ 632.055303] env[62066]: DEBUG nova.compute.manager [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 632.056032] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decd9105-ab35-4601-89c4-90478eb758f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.071720] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 632.071981] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating directory with path [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 632.072189] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-112a1639-677b-4606-b553-ed0fccc79c8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.102355] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Created directory with path [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 632.102486] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Fetch image to [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 632.102657] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Downloading image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk on the data store datastore1 {{(pid=62066) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 632.104063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e00adc0-5767-4725-a5ca-681f40625989 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.111785] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a3cedc-4f78-4023-8297-18c354f57798 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.122542] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354cfbbc-344f-42f6-9f77-0c4ccfa8907e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.153489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9c3b81-72a6-47e0-819e-a1477997caac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.159236] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4b0167f6-dbd7-403f-8ca7-114b3d0c1212 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.179709] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Downloading image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to the data store datastore1 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 632.231738] env[62066]: DEBUG oslo_vmware.rw_handles [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 632.287418] env[62066]: INFO nova.compute.manager [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] [instance: 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1] Took 1.07 seconds to deallocate network for instance. [ 632.396517] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7543869-9b8e-4842-bed2-e2a84d442d92 req-e7a90246-d487-4911-9496-e5fa41003c33 service nova] Releasing lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.396826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquired lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.397045] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.427500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.471545] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.472090] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.480054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.807s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.480054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.480054] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 632.480054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.482712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de39c53b-96a6-411b-a149-a190a5ef2fc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.494788] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c100d2-a87c-41a0-834f-5dcc1c6faba6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.512161] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcc8de8-4b00-47c1-8773-7f64ad129ebe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.523223] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6731cdaa-37dc-40ba-8911-f07ce6930f52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.560111] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181462MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 632.560303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.568993] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 632.569313] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58053b17-1f5a-41bd-9f44-0bd32e41f4cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.576068] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 632.576068] env[62066]: value = "task-1155698" [ 632.576068] env[62066]: _type = "Task" [ 632.576068] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.584614] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.876895] env[62066]: DEBUG oslo_vmware.rw_handles [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 632.877289] env[62066]: DEBUG oslo_vmware.rw_handles [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 632.916294] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.987092] env[62066]: DEBUG nova.compute.utils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 632.990180] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.990465] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 632.997600] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.020436] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Downloaded image file data fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk on the data store datastore1 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 633.022550] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 633.022791] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copying Virtual Disk [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk to [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 633.024026] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5768cac-7dbd-4e31-804d-02b032692f19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.037830] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 633.037830] env[62066]: value = "task-1155699" [ 633.037830] env[62066]: _type = "Task" [ 633.037830] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.048197] env[62066]: DEBUG nova.policy [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53366fea3a6434fadadede81df16089', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33b756ddd30f4cb0b917fad171213266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.055660] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.089914] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155698, 'name': PowerOffVM_Task, 'duration_secs': 0.184017} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.092677] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 633.092998] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.093984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59d4d8a-df31-42dc-b040-4fc804be72ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.100805] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 633.104385] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-660a8960-3b90-4b10-8672-c127aae71ce3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.132013] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 633.132932] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 633.133179] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Deleting the datastore file [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 633.133447] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-773cc015-748c-4879-a7fb-c7d73d3a5cb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.146186] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 633.146186] env[62066]: value = "task-1155701" [ 633.146186] env[62066]: _type = "Task" [ 633.146186] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.155042] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.322468] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Successfully created port: a18da785-d71f-443d-997d-a3aa7f5c6797 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.329208] env[62066]: INFO nova.scheduler.client.report [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Deleted allocations for instance 80fd9dcf-4094-4c96-aef5-f59fbde4e2b1 [ 633.375873] env[62066]: DEBUG nova.compute.manager [req-09ad074c-4acc-4e70-b6bb-ac6b7399bb4d req-786ac80f-26ba-474f-bc8f-8578e837dcfa service nova] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Received event network-vif-deleted-d0da24d6-830a-484a-af48-e42459fd3b21 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.418817] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47417c43-7884-44ec-8ce5-910a295eb015 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.427185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96afef3b-6c78-4375-94e7-123959d44129 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.459964] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7330c4-bc82-418e-8bb7-25472e8f0a42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.467988] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3210899-4af6-4e63-b789-9d9a961909ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.481942] env[62066]: DEBUG nova.compute.provider_tree [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.494222] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.500666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Releasing lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.501139] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.501271] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.501800] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46a2519b-41e9-4c91-a37c-5ec28b2a41a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.511186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60af2ed8-e78f-430d-bc9c-a5f21d42b6a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.535186] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 58528227-a9dc-4c12-b15b-fa191bc29046 could not be found. [ 633.535449] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.535634] env[62066]: INFO nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.535922] env[62066]: DEBUG oslo.service.loopingcall [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.536176] env[62066]: DEBUG nova.compute.manager [-] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.536267] env[62066]: DEBUG nova.network.neutron [-] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.550122] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155699, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.557570] env[62066]: DEBUG nova.network.neutron [-] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.657749] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082752} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.658028] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 633.658470] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 633.658659] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.852387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef24993b-80d2-43c5-995d-e46b7031c04f tempest-ServerMetadataTestJSON-2025188711 tempest-ServerMetadataTestJSON-2025188711-project-member] Lock "80fd9dcf-4094-4c96-aef5-f59fbde4e2b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.392s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.988695] env[62066]: DEBUG nova.scheduler.client.report [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.056907] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155699, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.060124] env[62066]: DEBUG nova.network.neutron [-] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.359749] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.496407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.497242] env[62066]: ERROR nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Traceback (most recent call last): [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.driver.spawn(context, instance, image_meta, [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] vm_ref = self.build_virtual_machine(instance, [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.497242] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] for vif in network_info: [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return self._sync_wrapper(fn, *args, **kwargs) [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.wait() [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self[:] = self._gt.wait() [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return self._exit_event.wait() [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] current.throw(*self._exc) [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.497608] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] result = function(*args, **kwargs) [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] return func(*args, **kwargs) [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise e [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] nwinfo = self.network_api.allocate_for_instance( [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] created_port_ids = self._update_ports_for_instance( [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] with excutils.save_and_reraise_exception(): [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] self.force_reraise() [ 634.497920] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise self.value [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] updated_port = self._update_port( [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] _ensure_no_port_binding_failure(port) [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] raise exception.PortBindingFailed(port_id=port['id']) [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] nova.exception.PortBindingFailed: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. [ 634.498273] env[62066]: ERROR nova.compute.manager [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] [ 634.498273] env[62066]: DEBUG nova.compute.utils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 634.499445] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.101s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.501602] env[62066]: INFO nova.compute.claims [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.504594] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Build of instance b5610b23-828b-4ece-b18a-4bfe27a60bbe was re-scheduled: Binding failed for port a56380e7-deee-408b-9b19-5661ee095742, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.505040] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.505267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.505498] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.505665] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.507295] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.547752] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.548421] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.548812] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.549208] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.549509] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.549803] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.550038] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.550301] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.550488] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.550650] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.550818] env[62066]: DEBUG nova.virt.hardware [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.554043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d6527e-3a26-4340-8aef-c8548cb6f562 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.558017] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155699, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.562604] env[62066]: INFO nova.compute.manager [-] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Took 1.03 seconds to deallocate network for instance. [ 634.565344] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe76744-896a-4693-b8d6-594af6d8fd0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.569333] env[62066]: DEBUG nova.compute.claims [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.569454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.693908] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.694401] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.697948] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.697948] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.697948] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.697948] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.697948] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.698113] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.698113] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.698113] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.698113] env[62066]: DEBUG nova.virt.hardware [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.698113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bc9bb2-8bce-45e5-8370-9326fcf7bb0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.710634] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab37d402-bbaf-4e6d-b7d0-09a0ebba2e4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.727614] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.734384] env[62066]: DEBUG oslo.service.loopingcall [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.734541] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.734759] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab7cf327-f7d6-4378-bff7-7cdd172f584c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.753127] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.753127] env[62066]: value = "task-1155702" [ 634.753127] env[62066]: _type = "Task" [ 634.753127] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.762069] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155702, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.895025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.947874] env[62066]: ERROR nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 634.947874] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.947874] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.947874] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.947874] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.947874] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.947874] env[62066]: ERROR nova.compute.manager raise self.value [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.947874] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.947874] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.947874] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.948369] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.948369] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.948369] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 634.948369] env[62066]: ERROR nova.compute.manager [ 634.948369] env[62066]: Traceback (most recent call last): [ 634.948369] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.948369] env[62066]: listener.cb(fileno) [ 634.948369] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.948369] env[62066]: result = function(*args, **kwargs) [ 634.948369] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.948369] env[62066]: return func(*args, **kwargs) [ 634.948369] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.948369] env[62066]: raise e [ 634.948369] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.948369] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 634.948369] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.948369] env[62066]: created_port_ids = self._update_ports_for_instance( [ 634.948369] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.948369] env[62066]: with excutils.save_and_reraise_exception(): [ 634.948369] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.948369] env[62066]: self.force_reraise() [ 634.948369] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.948369] env[62066]: raise self.value [ 634.948369] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.948369] env[62066]: updated_port = self._update_port( [ 634.948369] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.948369] env[62066]: _ensure_no_port_binding_failure(port) [ 634.948369] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.948369] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.949073] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 634.949073] env[62066]: Removing descriptor: 18 [ 634.949073] env[62066]: ERROR nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Traceback (most recent call last): [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] yield resources [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.driver.spawn(context, instance, image_meta, [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.949073] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] vm_ref = self.build_virtual_machine(instance, [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] for vif in network_info: [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self._sync_wrapper(fn, *args, **kwargs) [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.wait() [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self[:] = self._gt.wait() [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self._exit_event.wait() [ 634.949369] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] result = hub.switch() [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self.greenlet.switch() [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] result = function(*args, **kwargs) [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return func(*args, **kwargs) [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise e [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] nwinfo = self.network_api.allocate_for_instance( [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.949689] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] created_port_ids = self._update_ports_for_instance( [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] with excutils.save_and_reraise_exception(): [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.force_reraise() [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise self.value [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] updated_port = self._update_port( [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] _ensure_no_port_binding_failure(port) [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.950053] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise exception.PortBindingFailed(port_id=port['id']) [ 634.950382] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 634.950382] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] [ 634.950382] env[62066]: INFO nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Terminating instance [ 634.951731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.951892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.952069] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.040531] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.056723] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155699, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.650923} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.057077] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copied Virtual Disk [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk to [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.057299] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleting the datastore file [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/tmp-sparse.vmdk {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 635.057602] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fd2466c-631f-4e70-9372-363b51dbb397 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.065577] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 635.065577] env[62066]: value = "task-1155703" [ 635.065577] env[62066]: _type = "Task" [ 635.065577] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.074892] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.265704] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155702, 'name': CreateVM_Task, 'duration_secs': 0.303686} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.266081] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 635.266367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.266548] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.266910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 635.267265] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b83090a1-dc77-4cc3-80f9-741e335c5910 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.275184] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 635.275184] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52042f11-7f8a-25f7-bed7-f5b37d175fc8" [ 635.275184] env[62066]: _type = "Task" [ 635.275184] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.283367] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52042f11-7f8a-25f7-bed7-f5b37d175fc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.405084] env[62066]: DEBUG nova.compute.manager [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Received event network-changed-a18da785-d71f-443d-997d-a3aa7f5c6797 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.406304] env[62066]: DEBUG nova.compute.manager [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Refreshing instance network info cache due to event network-changed-a18da785-d71f-443d-997d-a3aa7f5c6797. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 635.406304] env[62066]: DEBUG oslo_concurrency.lockutils [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] Acquiring lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.409427] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.496895] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.585358] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.587540] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 635.587914] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Moving file from [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df/fc5145ed-66bc-4490-b8ac-7ca0de814dc0 to [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0. {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 635.589481] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-b1169c87-b6dc-4f74-a323-0db7a82623d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.602046] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 635.602046] env[62066]: value = "task-1155704" [ 635.602046] env[62066]: _type = "Task" [ 635.602046] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.619170] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155704, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.676340] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.787728] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52042f11-7f8a-25f7-bed7-f5b37d175fc8, 'name': SearchDatastore_Task, 'duration_secs': 0.042575} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.787728] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.787728] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.787907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.787907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.789207] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.789207] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-467b8ea0-e6af-4643-a345-cd1b340b892e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.799660] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.799886] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.800877] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1860811-265f-4b98-9e77-45b75673dd3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.806656] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 635.806656] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5208dc31-3499-caca-1e10-acc1337d94f3" [ 635.806656] env[62066]: _type = "Task" [ 635.806656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.814648] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5208dc31-3499-caca-1e10-acc1337d94f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.911080] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "refresh_cache-b5610b23-828b-4ece-b18a-4bfe27a60bbe" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.911664] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 635.912059] env[62066]: DEBUG nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.912110] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.937174] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.991014] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a780a86f-1fe1-4c55-b01b-a03927d45669 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.000105] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8364315-5c2a-4f01-8f99-9b6b7064b6bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.036535] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3885aa3e-b1e2-4e3a-8854-76849661fe9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.044863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a619d436-e914-42c3-8f65-62b54893ada8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.064975] env[62066]: DEBUG nova.compute.provider_tree [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.114359] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155704, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030094} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.114623] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] File moved {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 636.114824] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Cleaning up location [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 636.114975] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleting the datastore file [datastore1] vmware_temp/184fd652-ca6d-49f3-b630-787be07c46df {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 636.115281] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-051810f7-def3-4254-a13a-572adf4cb3f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.122702] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 636.122702] env[62066]: value = "task-1155705" [ 636.122702] env[62066]: _type = "Task" [ 636.122702] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.131264] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.181945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.182954] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.183313] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.183950] env[62066]: DEBUG oslo_concurrency.lockutils [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] Acquired lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.184249] env[62066]: DEBUG nova.network.neutron [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Refreshing network info cache for port a18da785-d71f-443d-997d-a3aa7f5c6797 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.185927] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca19670-9188-401c-a39c-3c63d41fcd3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.198792] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd20b8bf-2658-4f86-ac8c-ef7e21d79c86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.224621] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a406e14c-b281-4d1f-97d0-3219c6ffaa8d could not be found. [ 636.224691] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.224840] env[62066]: INFO nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.225094] env[62066]: DEBUG oslo.service.loopingcall [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.225333] env[62066]: DEBUG nova.compute.manager [-] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.225428] env[62066]: DEBUG nova.network.neutron [-] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.241831] env[62066]: DEBUG nova.network.neutron [-] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.320297] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5208dc31-3499-caca-1e10-acc1337d94f3, 'name': SearchDatastore_Task, 'duration_secs': 0.00842} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.321583] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f0b3f8f-7f28-435c-afe9-8a42c55621a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.328509] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 636.328509] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529fa44f-8297-94bd-6f8f-bedc319328f4" [ 636.328509] env[62066]: _type = "Task" [ 636.328509] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.336681] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529fa44f-8297-94bd-6f8f-bedc319328f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.440065] env[62066]: DEBUG nova.network.neutron [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.570208] env[62066]: DEBUG nova.scheduler.client.report [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.634048] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024896} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.634325] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 636.635051] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b4593d-6b5d-48bd-85f9-b8e6aa662a97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.641883] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 636.641883] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e67c5e-44b7-2172-8d21-86e250f29358" [ 636.641883] env[62066]: _type = "Task" [ 636.641883] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.650210] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e67c5e-44b7-2172-8d21-86e250f29358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.717774] env[62066]: DEBUG nova.network.neutron [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.749913] env[62066]: DEBUG nova.network.neutron [-] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.842819] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529fa44f-8297-94bd-6f8f-bedc319328f4, 'name': SearchDatastore_Task, 'duration_secs': 0.008763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.843295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.843718] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 636.844570] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cc0d82d-e450-4f0e-bb64-9d1214522307 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.851717] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 636.851717] env[62066]: value = "task-1155706" [ 636.851717] env[62066]: _type = "Task" [ 636.851717] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.864066] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.914487] env[62066]: DEBUG nova.network.neutron [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.945017] env[62066]: INFO nova.compute.manager [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: b5610b23-828b-4ece-b18a-4bfe27a60bbe] Took 1.03 seconds to deallocate network for instance. [ 637.079730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.080511] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.084140] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.345s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.156082] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e67c5e-44b7-2172-8d21-86e250f29358, 'name': SearchDatastore_Task, 'duration_secs': 0.008554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.156488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.156877] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 637.157212] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2cdcd7b-5bd9-43f2-a6b7-b554a9d76526 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.167023] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 637.167023] env[62066]: value = "task-1155707" [ 637.167023] env[62066]: _type = "Task" [ 637.167023] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.178620] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.254356] env[62066]: INFO nova.compute.manager [-] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Took 1.03 seconds to deallocate network for instance. [ 637.257386] env[62066]: DEBUG nova.compute.claims [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.257678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.363528] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.363812] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 637.364027] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 637.364464] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86b560da-d091-46e5-bb91-be5f0cd5c6ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.371209] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 637.371209] env[62066]: value = "task-1155708" [ 637.371209] env[62066]: _type = "Task" [ 637.371209] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.378927] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.417585] env[62066]: DEBUG oslo_concurrency.lockutils [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] Releasing lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.417701] env[62066]: DEBUG nova.compute.manager [req-b362984c-f815-48f2-99b3-345b909811f4 req-f9c24143-a3c6-48af-ad52-c999a9d19a4a service nova] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Received event network-vif-deleted-a18da785-d71f-443d-997d-a3aa7f5c6797 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.589656] env[62066]: DEBUG nova.compute.utils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 637.595597] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.595597] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 637.642877] env[62066]: DEBUG nova.policy [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c378752334ce478da517eb58c2d31e00', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd191018b67d416bb508b7754f141525', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.677956] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155707, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.885083] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.961016] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Successfully created port: 3c6574e9-8721-409b-846a-39cf31c97460 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.979529] env[62066]: INFO nova.scheduler.client.report [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Deleted allocations for instance b5610b23-828b-4ece-b18a-4bfe27a60bbe [ 638.019712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b07573-f4c9-4be9-9d1f-dc9f7727cd45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.029228] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4a49bc-5c06-4062-af5b-11afa8592ebe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.062819] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847b4107-b030-416c-b7fd-9c5028eb3cc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.070480] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bba9687-204c-4f74-923a-88df1cba02fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.085446] env[62066]: DEBUG nova.compute.provider_tree [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.098704] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.176430] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543859} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.177013] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 638.177245] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.177493] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee36442d-32a0-44eb-8632-4475a135bb50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.184532] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 638.184532] env[62066]: value = "task-1155709" [ 638.184532] env[62066]: _type = "Task" [ 638.184532] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.192409] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.382642] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.879146} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.382920] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.383709] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfda6e2-17a5-472c-8d6b-2b2d1cc8bbde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.404246] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.404918] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1ac8008-ae09-4d0f-94af-1e6d58a36de7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.425121] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 638.425121] env[62066]: value = "task-1155710" [ 638.425121] env[62066]: _type = "Task" [ 638.425121] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.433416] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155710, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.492523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5196e88-aaee-490a-ae30-a9200c2404e5 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "b5610b23-828b-4ece-b18a-4bfe27a60bbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.515s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.497790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.498023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.589308] env[62066]: DEBUG nova.scheduler.client.report [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.694861] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.21821} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.694861] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.697104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6437b8cf-85be-4698-9348-c6cfafcbc172 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.717111] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.718700] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-358ce037-f37d-4fbe-950d-55ca1ad96e32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.735784] env[62066]: DEBUG nova.compute.manager [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Received event network-changed-3c6574e9-8721-409b-846a-39cf31c97460 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.735994] env[62066]: DEBUG nova.compute.manager [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Refreshing instance network info cache due to event network-changed-3c6574e9-8721-409b-846a-39cf31c97460. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 638.736503] env[62066]: DEBUG oslo_concurrency.lockutils [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] Acquiring lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.736661] env[62066]: DEBUG oslo_concurrency.lockutils [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] Acquired lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.736826] env[62066]: DEBUG nova.network.neutron [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Refreshing network info cache for port 3c6574e9-8721-409b-846a-39cf31c97460 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 638.744183] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 638.744183] env[62066]: value = "task-1155711" [ 638.744183] env[62066]: _type = "Task" [ 638.744183] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.753580] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155711, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.885849] env[62066]: ERROR nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 638.885849] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.885849] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.885849] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.885849] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.885849] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.885849] env[62066]: ERROR nova.compute.manager raise self.value [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.885849] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.885849] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.885849] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.886371] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.886371] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.886371] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 638.886371] env[62066]: ERROR nova.compute.manager [ 638.886371] env[62066]: Traceback (most recent call last): [ 638.886371] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.886371] env[62066]: listener.cb(fileno) [ 638.886371] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.886371] env[62066]: result = function(*args, **kwargs) [ 638.886371] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.886371] env[62066]: return func(*args, **kwargs) [ 638.886371] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.886371] env[62066]: raise e [ 638.886371] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.886371] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 638.886371] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.886371] env[62066]: created_port_ids = self._update_ports_for_instance( [ 638.886371] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.886371] env[62066]: with excutils.save_and_reraise_exception(): [ 638.886371] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.886371] env[62066]: self.force_reraise() [ 638.886371] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.886371] env[62066]: raise self.value [ 638.886371] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.886371] env[62066]: updated_port = self._update_port( [ 638.886371] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.886371] env[62066]: _ensure_no_port_binding_failure(port) [ 638.886371] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.886371] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.887179] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 638.887179] env[62066]: Removing descriptor: 18 [ 638.934705] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.999159] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.097882] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.098594] env[62066]: ERROR nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Traceback (most recent call last): [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.driver.spawn(context, instance, image_meta, [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] vm_ref = self.build_virtual_machine(instance, [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.098594] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] for vif in network_info: [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return self._sync_wrapper(fn, *args, **kwargs) [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.wait() [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self[:] = self._gt.wait() [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return self._exit_event.wait() [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] current.throw(*self._exc) [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.098996] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] result = function(*args, **kwargs) [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] return func(*args, **kwargs) [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise e [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] nwinfo = self.network_api.allocate_for_instance( [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] created_port_ids = self._update_ports_for_instance( [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] with excutils.save_and_reraise_exception(): [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] self.force_reraise() [ 639.100652] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise self.value [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] updated_port = self._update_port( [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] _ensure_no_port_binding_failure(port) [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] raise exception.PortBindingFailed(port_id=port['id']) [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] nova.exception.PortBindingFailed: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. [ 639.101718] env[62066]: ERROR nova.compute.manager [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] [ 639.101718] env[62066]: DEBUG nova.compute.utils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 639.102144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.378s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.102389] env[62066]: INFO nova.compute.claims [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 639.105167] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Build of instance 5fa586bb-bf1c-4126-a2a7-70e488307794 was re-scheduled: Binding failed for port 3708b533-8f03-4217-a772-aecc26bd1bcc, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.106079] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 639.106079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquiring lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.106079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Acquired lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.106334] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.108240] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.135077] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.135342] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.135630] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.135712] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.135858] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.136035] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.136258] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.136427] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.136601] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.137202] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.137202] env[62066]: DEBUG nova.virt.hardware [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.138095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c7415d-3598-42c9-b868-d94f22db69c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.146394] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0b2ee4-8741-4389-aa27-55815e83d6b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.161137] env[62066]: ERROR nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Traceback (most recent call last): [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] yield resources [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.driver.spawn(context, instance, image_meta, [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] vm_ref = self.build_virtual_machine(instance, [ 639.161137] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] for vif in network_info: [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return self._sync_wrapper(fn, *args, **kwargs) [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.wait() [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self[:] = self._gt.wait() [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return self._exit_event.wait() [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.161485] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] current.throw(*self._exc) [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] result = function(*args, **kwargs) [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return func(*args, **kwargs) [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise e [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] nwinfo = self.network_api.allocate_for_instance( [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] created_port_ids = self._update_ports_for_instance( [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] with excutils.save_and_reraise_exception(): [ 639.161815] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.force_reraise() [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise self.value [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] updated_port = self._update_port( [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] _ensure_no_port_binding_failure(port) [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise exception.PortBindingFailed(port_id=port['id']) [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 639.162142] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] [ 639.162142] env[62066]: INFO nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Terminating instance [ 639.166733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquiring lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.253936] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155711, 'name': ReconfigVM_Task, 'duration_secs': 0.315189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.254252] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Reconfigured VM instance instance-00000019 to attach disk [datastore1] 2dc8d357-2864-410a-93ef-21ea4e9b530c/2dc8d357-2864-410a-93ef-21ea4e9b530c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.254852] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c754a1a6-26a6-4198-b9d5-71208df8a848 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.256985] env[62066]: DEBUG nova.network.neutron [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.263687] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 639.263687] env[62066]: value = "task-1155712" [ 639.263687] env[62066]: _type = "Task" [ 639.263687] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.271886] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155712, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.342423] env[62066]: DEBUG nova.network.neutron [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.438268] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.519716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.626246] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.698280] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.774278] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155712, 'name': Rename_Task, 'duration_secs': 0.135663} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.776220] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 639.776220] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74f82eb3-5db7-4fa6-81f5-0f1e0eb88638 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.781146] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 639.781146] env[62066]: value = "task-1155713" [ 639.781146] env[62066]: _type = "Task" [ 639.781146] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.789007] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.845824] env[62066]: DEBUG oslo_concurrency.lockutils [req-90a69e89-1d46-46be-9976-7e9c5388f2e4 req-b3246c7d-847c-40af-9800-0c64bdfdf54a service nova] Releasing lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.846592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquired lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.847250] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.935781] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155710, 'name': ReconfigVM_Task, 'duration_secs': 1.218019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.936106] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfigured VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 639.936718] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fd0d36b-541b-4f39-8deb-f7e7f3e91221 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.943516] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 639.943516] env[62066]: value = "task-1155714" [ 639.943516] env[62066]: _type = "Task" [ 639.943516] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.951360] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155714, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.200284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Releasing lock "refresh_cache-5fa586bb-bf1c-4126-a2a7-70e488307794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.200598] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 640.200783] env[62066]: DEBUG nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.200951] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.219595] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.294059] env[62066]: DEBUG oslo_vmware.api [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155713, 'name': PowerOnVM_Task, 'duration_secs': 0.452778} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.294342] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.294562] env[62066]: DEBUG nova.compute.manager [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 640.295315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76056564-ea33-4e86-a58d-0e8d8aa1ee66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.365307] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.455513] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155714, 'name': Rename_Task, 'duration_secs': 0.12453} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.457999] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 640.458490] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7c6b515-c500-444f-8675-4b1d532ec8c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.464943] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 640.464943] env[62066]: value = "task-1155715" [ 640.464943] env[62066]: _type = "Task" [ 640.464943] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.474841] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.482374] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155715, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.485155] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8172d656-c98e-47ae-9ec5-6ad55ae6acb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.492117] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dec81d-a410-4549-92ed-acb30f673d9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.525262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf92c37-ac59-47ce-8d32-4d8284971bdc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.533162] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f3a357-8f82-460a-b5e2-6ea3ae1bf7a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.549212] env[62066]: DEBUG nova.compute.provider_tree [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.727598] env[62066]: DEBUG nova.network.neutron [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.814659] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.819791] env[62066]: DEBUG nova.compute.manager [req-f1916d2f-5a01-4f6b-af09-35066772c73a req-f93a3f5b-be61-40f2-973e-cb4c723213ca service nova] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Received event network-vif-deleted-3c6574e9-8721-409b-846a-39cf31c97460 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.974849] env[62066]: DEBUG oslo_vmware.api [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155715, 'name': PowerOnVM_Task, 'duration_secs': 0.426061} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.975123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.975326] env[62066]: DEBUG nova.compute.manager [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 640.976109] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bf1b60-93d1-4c53-8c3d-4ceb996fa5f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.984049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Releasing lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.984427] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.984618] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.985052] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60e4a000-673c-4280-b430-5a1f6b37fc82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.992832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebf1e04-e327-4fdc-a599-f362b54f018a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.017598] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d347bf40-941c-428f-85c3-18067a21d6c9 could not be found. [ 641.017848] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.018039] env[62066]: INFO nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 641.018336] env[62066]: DEBUG oslo.service.loopingcall [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.019137] env[62066]: DEBUG nova.compute.manager [-] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.019256] env[62066]: DEBUG nova.network.neutron [-] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.034977] env[62066]: DEBUG nova.network.neutron [-] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.051555] env[62066]: DEBUG nova.scheduler.client.report [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.230920] env[62066]: INFO nova.compute.manager [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] [instance: 5fa586bb-bf1c-4126-a2a7-70e488307794] Took 1.03 seconds to deallocate network for instance. [ 641.385117] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.385407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.385672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "2dc8d357-2864-410a-93ef-21ea4e9b530c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.385821] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.385990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.388291] env[62066]: INFO nova.compute.manager [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Terminating instance [ 641.390065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "refresh_cache-2dc8d357-2864-410a-93ef-21ea4e9b530c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.390229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "refresh_cache-2dc8d357-2864-410a-93ef-21ea4e9b530c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.390395] env[62066]: DEBUG nova.network.neutron [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.493355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.537372] env[62066]: DEBUG nova.network.neutron [-] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.556240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.556758] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.559973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.161s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.561445] env[62066]: INFO nova.compute.claims [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.907794] env[62066]: DEBUG nova.network.neutron [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.958909] env[62066]: DEBUG nova.network.neutron [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.040110] env[62066]: INFO nova.compute.manager [-] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Took 1.02 seconds to deallocate network for instance. [ 642.041952] env[62066]: DEBUG nova.compute.claims [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.042293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.066014] env[62066]: DEBUG nova.compute.utils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 642.069778] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 642.069778] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 642.128508] env[62066]: DEBUG nova.policy [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5caad5da140d457d88c40c304dd4f605', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab8a3b607934456b5d408442f03524a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 642.267819] env[62066]: INFO nova.scheduler.client.report [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Deleted allocations for instance 5fa586bb-bf1c-4126-a2a7-70e488307794 [ 642.427427] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Successfully created port: dd3173d0-dd91-44a1-a785-43c57a362fbf {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.461607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "refresh_cache-2dc8d357-2864-410a-93ef-21ea4e9b530c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.462089] env[62066]: DEBUG nova.compute.manager [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.462255] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.463246] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0acbe29-3e56-4ef8-a9bf-b3fd3e096ebd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.471525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 642.471775] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9f30ebc-194b-4380-b835-6fc345bd3998 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.478799] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 642.478799] env[62066]: value = "task-1155716" [ 642.478799] env[62066]: _type = "Task" [ 642.478799] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.487996] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155716, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.506155] env[62066]: INFO nova.compute.manager [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Rebuilding instance [ 642.544587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "641bf537-76d2-4c41-a6dc-b5b042327477" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.544861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.554837] env[62066]: DEBUG nova.compute.manager [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 642.555785] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7da5748-0b91-4d15-8866-d1241c66a334 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.572083] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 642.780152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1051205d-8e90-4b3b-b784-15ab9d6826d2 tempest-VolumesAdminNegativeTest-777943893 tempest-VolumesAdminNegativeTest-777943893-project-member] Lock "5fa586bb-bf1c-4126-a2a7-70e488307794" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.990357] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800a5364-90bb-407f-b8a4-5a494f83f42c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.995792] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155716, 'name': PowerOffVM_Task, 'duration_secs': 0.301816} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.996473] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 642.996907] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 642.996982] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-875ef4a5-1b25-48e2-acbf-f77c3e91fe85 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.001465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff670a0-a6c5-4fb9-bab6-2d3b973acfd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.038917] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef78fb3-7d70-43ad-b3e9-dcb15e2d60de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.040046] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.040240] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.040403] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleting the datastore file [datastore1] 2dc8d357-2864-410a-93ef-21ea4e9b530c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.040700] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63898178-e60f-4121-8296-5cd2bf3d4277 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.048988] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815a46d1-38ec-4659-b56d-5e609f79f77e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.054780] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 643.054780] env[62066]: value = "task-1155718" [ 643.054780] env[62066]: _type = "Task" [ 643.054780] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.071969] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 643.072533] env[62066]: DEBUG nova.compute.provider_tree [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.074076] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b5b156c-181d-439a-98cf-df748074896f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.083632] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.088986] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 643.088986] env[62066]: value = "task-1155719" [ 643.088986] env[62066]: _type = "Task" [ 643.088986] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.099500] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.160476] env[62066]: DEBUG nova.compute.manager [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Received event network-changed-dd3173d0-dd91-44a1-a785-43c57a362fbf {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.160694] env[62066]: DEBUG nova.compute.manager [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Refreshing instance network info cache due to event network-changed-dd3173d0-dd91-44a1-a785-43c57a362fbf. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 643.160940] env[62066]: DEBUG oslo_concurrency.lockutils [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] Acquiring lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.161142] env[62066]: DEBUG oslo_concurrency.lockutils [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] Acquired lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.161335] env[62066]: DEBUG nova.network.neutron [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Refreshing network info cache for port dd3173d0-dd91-44a1-a785-43c57a362fbf {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 643.283202] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.347136] env[62066]: ERROR nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 643.347136] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.347136] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.347136] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.347136] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.347136] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.347136] env[62066]: ERROR nova.compute.manager raise self.value [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.347136] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 643.347136] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.347136] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 643.347861] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.347861] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 643.347861] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 643.347861] env[62066]: ERROR nova.compute.manager [ 643.347861] env[62066]: Traceback (most recent call last): [ 643.347861] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 643.347861] env[62066]: listener.cb(fileno) [ 643.347861] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.347861] env[62066]: result = function(*args, **kwargs) [ 643.347861] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.347861] env[62066]: return func(*args, **kwargs) [ 643.347861] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.347861] env[62066]: raise e [ 643.347861] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.347861] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 643.347861] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.347861] env[62066]: created_port_ids = self._update_ports_for_instance( [ 643.347861] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.347861] env[62066]: with excutils.save_and_reraise_exception(): [ 643.347861] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.347861] env[62066]: self.force_reraise() [ 643.347861] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.347861] env[62066]: raise self.value [ 643.347861] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.347861] env[62066]: updated_port = self._update_port( [ 643.347861] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.347861] env[62066]: _ensure_no_port_binding_failure(port) [ 643.347861] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.347861] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 643.348848] env[62066]: nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 643.348848] env[62066]: Removing descriptor: 18 [ 643.565454] env[62066]: DEBUG oslo_vmware.api [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118057} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.565787] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.566016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 643.566248] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.566516] env[62066]: INFO nova.compute.manager [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 643.566784] env[62066]: DEBUG oslo.service.loopingcall [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.567010] env[62066]: DEBUG nova.compute.manager [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.567142] env[62066]: DEBUG nova.network.neutron [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.577807] env[62066]: DEBUG nova.scheduler.client.report [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.585294] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.587817] env[62066]: DEBUG nova.network.neutron [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.597843] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155719, 'name': PowerOffVM_Task, 'duration_secs': 0.177031} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.598918] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.599160] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.599951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b93a6a-4303-4e71-93f8-269d61c463bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.608481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 643.608753] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5b9dc95-6e57-4f0d-87de-35fa8d44d92f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.612110] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.612375] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.612582] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.612805] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.612983] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.613186] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.613425] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.613617] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.613813] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.614013] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.614223] env[62066]: DEBUG nova.virt.hardware [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.615249] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c46a577-37bd-46bd-b4e8-dbb93f592856 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.623262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d8c3ca-92b7-4aaa-9f82-e31d61fd308a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.637455] env[62066]: ERROR nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Traceback (most recent call last): [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] yield resources [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.driver.spawn(context, instance, image_meta, [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] vm_ref = self.build_virtual_machine(instance, [ 643.637455] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] for vif in network_info: [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return self._sync_wrapper(fn, *args, **kwargs) [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.wait() [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self[:] = self._gt.wait() [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return self._exit_event.wait() [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 643.637876] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] current.throw(*self._exc) [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] result = function(*args, **kwargs) [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return func(*args, **kwargs) [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise e [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] nwinfo = self.network_api.allocate_for_instance( [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] created_port_ids = self._update_ports_for_instance( [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] with excutils.save_and_reraise_exception(): [ 643.638298] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.force_reraise() [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise self.value [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] updated_port = self._update_port( [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] _ensure_no_port_binding_failure(port) [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise exception.PortBindingFailed(port_id=port['id']) [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 643.638664] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] [ 643.638664] env[62066]: INFO nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Terminating instance [ 643.640375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.641601] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.641840] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.642063] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Deleting the datastore file [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.642346] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f96fecf5-f02a-49b5-98d1-1d40579d61b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.648899] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 643.648899] env[62066]: value = "task-1155721" [ 643.648899] env[62066]: _type = "Task" [ 643.648899] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.658649] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.679402] env[62066]: DEBUG nova.network.neutron [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.755204] env[62066]: DEBUG nova.network.neutron [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.806190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.085939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.085939] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.087247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.936s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.089152] env[62066]: INFO nova.compute.claims [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.092622] env[62066]: DEBUG nova.network.neutron [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.162528] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080727} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.162528] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.162528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 644.162528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.257953] env[62066]: DEBUG oslo_concurrency.lockutils [req-4360cd0d-6e0b-46e7-abdb-3dae65218b9b req-93361045-54f6-44b2-899b-169f68407551 service nova] Releasing lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.258449] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.258636] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.594257] env[62066]: DEBUG nova.compute.utils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 644.600167] env[62066]: INFO nova.compute.manager [-] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Took 1.03 seconds to deallocate network for instance. [ 644.600167] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.600167] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 644.654674] env[62066]: DEBUG nova.policy [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fb2f95024cb4fb9a3dd044125d0bc97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26a6f97f438942ac8577b4a99edb7990', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.785739] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.925775] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.032435] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Successfully created port: cb97e1cb-5857-4b19-893b-5a360521e01c {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.105493] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.112452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.194825] env[62066]: DEBUG nova.compute.manager [req-0a087584-d7f6-4e89-ae65-d7e6b4d04a0c req-50fa3f3f-0e70-4765-b086-524f6e822dc6 service nova] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Received event network-vif-deleted-dd3173d0-dd91-44a1-a785-43c57a362fbf {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.211456] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.211706] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.211862] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.212056] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.212245] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.212383] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.212698] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.212864] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.213055] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.213231] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.213395] env[62066]: DEBUG nova.virt.hardware [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.214550] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689be792-dc3f-4070-b993-46323e806ec0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.225370] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de64da8-a5fc-4bc7-a0f1-b577fe95d360 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.240848] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 645.246635] env[62066]: DEBUG oslo.service.loopingcall [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.249366] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 645.249819] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d7e39e1-80b1-4fcf-b0c8-9dea2bb01be0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.270672] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 645.270672] env[62066]: value = "task-1155722" [ 645.270672] env[62066]: _type = "Task" [ 645.270672] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.279987] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155722, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.430611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.432016] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.432016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.432016] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bc558b0-c074-4477-a413-561d2c1379f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.448575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a06e92e-b805-4be7-84a7-890ff9ee2d34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.478397] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a943544f-28d3-4e3f-8d95-0b1cc6f2c59d could not be found. [ 645.478661] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.478881] env[62066]: INFO nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 645.479156] env[62066]: DEBUG oslo.service.loopingcall [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.482114] env[62066]: DEBUG nova.compute.manager [-] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.482218] env[62066]: DEBUG nova.network.neutron [-] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.499859] env[62066]: DEBUG nova.network.neutron [-] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.576314] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586de6aa-d109-4370-8360-9cc6e6af6632 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.583079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764a9dde-560c-486e-a1bc-41c7bd893d9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.619061] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19a8ec6-5dc4-40cc-8177-ca4879c5f6d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.626971] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ea29e5-21b9-498f-a610-6fad4b1b7c26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.642452] env[62066]: DEBUG nova.compute.provider_tree [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.791943] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155722, 'name': CreateVM_Task, 'duration_secs': 0.279649} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.792405] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 645.792990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.793507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.793507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 645.793799] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ada68223-f75a-4646-a8e0-e779cfa43c62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.799856] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 645.799856] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52169619-cf25-6d9e-e950-a5576060e8ba" [ 645.799856] env[62066]: _type = "Task" [ 645.799856] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.807736] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52169619-cf25-6d9e-e950-a5576060e8ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.928109] env[62066]: ERROR nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 645.928109] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.928109] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.928109] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.928109] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.928109] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.928109] env[62066]: ERROR nova.compute.manager raise self.value [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.928109] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.928109] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.928109] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.928547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.928547] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.928547] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 645.928547] env[62066]: ERROR nova.compute.manager [ 645.928547] env[62066]: Traceback (most recent call last): [ 645.928547] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.928547] env[62066]: listener.cb(fileno) [ 645.928547] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.928547] env[62066]: result = function(*args, **kwargs) [ 645.928547] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.928547] env[62066]: return func(*args, **kwargs) [ 645.928547] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.928547] env[62066]: raise e [ 645.928547] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.928547] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 645.928547] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.928547] env[62066]: created_port_ids = self._update_ports_for_instance( [ 645.928547] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.928547] env[62066]: with excutils.save_and_reraise_exception(): [ 645.928547] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.928547] env[62066]: self.force_reraise() [ 645.928547] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.928547] env[62066]: raise self.value [ 645.928547] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.928547] env[62066]: updated_port = self._update_port( [ 645.928547] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.928547] env[62066]: _ensure_no_port_binding_failure(port) [ 645.928547] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.928547] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.929242] env[62066]: nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 645.929242] env[62066]: Removing descriptor: 18 [ 646.004078] env[62066]: DEBUG nova.network.neutron [-] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.126591] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.145758] env[62066]: DEBUG nova.scheduler.client.report [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.154126] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.154390] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.154543] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.154751] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.154899] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.155053] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.155254] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.155408] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.155654] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.155718] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.155879] env[62066]: DEBUG nova.virt.hardware [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.157794] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235773a3-a623-408d-8e8c-c4caead45ca9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.168865] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0c35b8-aa8a-46ac-aa16-b42345d31c2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.185313] env[62066]: ERROR nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Traceback (most recent call last): [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] yield resources [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.driver.spawn(context, instance, image_meta, [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] vm_ref = self.build_virtual_machine(instance, [ 646.185313] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] for vif in network_info: [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return self._sync_wrapper(fn, *args, **kwargs) [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.wait() [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self[:] = self._gt.wait() [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return self._exit_event.wait() [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.186356] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] current.throw(*self._exc) [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] result = function(*args, **kwargs) [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return func(*args, **kwargs) [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise e [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] nwinfo = self.network_api.allocate_for_instance( [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] created_port_ids = self._update_ports_for_instance( [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] with excutils.save_and_reraise_exception(): [ 646.186793] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.force_reraise() [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise self.value [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] updated_port = self._update_port( [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] _ensure_no_port_binding_failure(port) [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise exception.PortBindingFailed(port_id=port['id']) [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 646.187170] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] [ 646.187170] env[62066]: INFO nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Terminating instance [ 646.190932] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquiring lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.191644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquired lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.191896] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.310091] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52169619-cf25-6d9e-e950-a5576060e8ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008163} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.310457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.310714] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 646.310943] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.311105] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.311282] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 646.311541] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f90103bc-fdfe-4d59-b1d3-a86dc32ec859 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.319034] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 646.319203] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 646.319995] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8385a9d-e288-40e6-95a2-8e0e95a86da8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.324601] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 646.324601] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d927ec-be7e-4faf-d297-84f30ad4554d" [ 646.324601] env[62066]: _type = "Task" [ 646.324601] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.332446] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d927ec-be7e-4faf-d297-84f30ad4554d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.506850] env[62066]: INFO nova.compute.manager [-] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Took 1.02 seconds to deallocate network for instance. [ 646.509489] env[62066]: DEBUG nova.compute.claims [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 646.509680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.651990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.652546] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.655771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.228s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.657079] env[62066]: INFO nova.compute.claims [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.717772] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.797731] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.835656] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d927ec-be7e-4faf-d297-84f30ad4554d, 'name': SearchDatastore_Task, 'duration_secs': 0.007382} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.836493] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-078aa9f9-d1d5-4b6c-b6c6-e39212f86599 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.841795] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 646.841795] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8ee30-c595-0f0e-a376-652d1ba55e1f" [ 646.841795] env[62066]: _type = "Task" [ 646.841795] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.850158] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8ee30-c595-0f0e-a376-652d1ba55e1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.164445] env[62066]: DEBUG nova.compute.utils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 647.166114] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.166313] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 647.222964] env[62066]: DEBUG nova.policy [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f0cb626a3c247de94d6931c8952cf44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e4a355991234435a0d6c31c7fc9d46c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.245706] env[62066]: DEBUG nova.compute.manager [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Received event network-changed-cb97e1cb-5857-4b19-893b-5a360521e01c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.245907] env[62066]: DEBUG nova.compute.manager [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Refreshing instance network info cache due to event network-changed-cb97e1cb-5857-4b19-893b-5a360521e01c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 647.246128] env[62066]: DEBUG oslo_concurrency.lockutils [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] Acquiring lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.304131] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Releasing lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.304131] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.304131] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.304131] env[62066]: DEBUG oslo_concurrency.lockutils [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] Acquired lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.304131] env[62066]: DEBUG nova.network.neutron [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Refreshing network info cache for port cb97e1cb-5857-4b19-893b-5a360521e01c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 647.304313] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88325dce-ef9b-4679-a069-4c5a890f0140 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.315129] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a480d2-6fcb-470c-85fb-3f03a8c97183 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.341823] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0b6deda-1870-44f3-b39a-e971d592be9e could not be found. [ 647.342116] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.342364] env[62066]: INFO nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 647.342630] env[62066]: DEBUG oslo.service.loopingcall [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.343145] env[62066]: DEBUG nova.compute.manager [-] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.343351] env[62066]: DEBUG nova.network.neutron [-] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.354919] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8ee30-c595-0f0e-a376-652d1ba55e1f, 'name': SearchDatastore_Task, 'duration_secs': 0.00854} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.355179] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.355422] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 647.355664] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-924df73e-3425-4518-a945-d9d3db91a1d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.362209] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 647.362209] env[62066]: value = "task-1155723" [ 647.362209] env[62066]: _type = "Task" [ 647.362209] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.366396] env[62066]: DEBUG nova.network.neutron [-] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.377216] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.500333] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Successfully created port: a8eed66c-0121-4e69-8d35-0afd4c422342 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.670539] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.831820] env[62066]: DEBUG nova.network.neutron [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.874999] env[62066]: DEBUG nova.network.neutron [-] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.882304] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155723, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465662} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.885307] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 647.885542] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 647.886732] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9c192cb-d951-4e29-8ae4-fd1ba95ca5ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.893891] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 647.893891] env[62066]: value = "task-1155724" [ 647.893891] env[62066]: _type = "Task" [ 647.893891] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.912585] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.914652] env[62066]: DEBUG nova.network.neutron [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.123645] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93eac34f-d82c-4770-a78e-d5af2157c7fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.133272] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06234d41-1bd9-49d7-bd89-c44f232e6d6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.167204] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94468aa-928d-42ea-ac0c-c633741cc899 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.178674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf98ebd-6487-4007-ae95-0a5068baffce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.193208] env[62066]: DEBUG nova.compute.provider_tree [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.384239] env[62066]: INFO nova.compute.manager [-] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Took 1.04 seconds to deallocate network for instance. [ 648.387013] env[62066]: DEBUG nova.compute.claims [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 648.387167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.402716] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062361} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.402956] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 648.403740] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83d1271-5454-47a7-97ad-b76aed28e479 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.423173] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 648.423638] env[62066]: DEBUG oslo_concurrency.lockutils [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] Releasing lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.423862] env[62066]: DEBUG nova.compute.manager [req-e59d151a-ca04-490c-974f-95f4d5dfd30a req-82ed5f34-a0c9-4fe7-93f9-756b38771b55 service nova] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Received event network-vif-deleted-cb97e1cb-5857-4b19-893b-5a360521e01c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.424176] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f85e248d-780c-4fe1-a7ae-ec58a0a20022 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.443831] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 648.443831] env[62066]: value = "task-1155725" [ 648.443831] env[62066]: _type = "Task" [ 648.443831] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.451907] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155725, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.460036] env[62066]: ERROR nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 648.460036] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.460036] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.460036] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.460036] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.460036] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.460036] env[62066]: ERROR nova.compute.manager raise self.value [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.460036] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.460036] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.460036] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.460638] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.460638] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.460638] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 648.460638] env[62066]: ERROR nova.compute.manager [ 648.460638] env[62066]: Traceback (most recent call last): [ 648.460638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.460638] env[62066]: listener.cb(fileno) [ 648.460638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.460638] env[62066]: result = function(*args, **kwargs) [ 648.460638] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.460638] env[62066]: return func(*args, **kwargs) [ 648.460638] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.460638] env[62066]: raise e [ 648.460638] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.460638] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 648.460638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.460638] env[62066]: created_port_ids = self._update_ports_for_instance( [ 648.460638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.460638] env[62066]: with excutils.save_and_reraise_exception(): [ 648.460638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.460638] env[62066]: self.force_reraise() [ 648.460638] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.460638] env[62066]: raise self.value [ 648.460638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.460638] env[62066]: updated_port = self._update_port( [ 648.460638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.460638] env[62066]: _ensure_no_port_binding_failure(port) [ 648.460638] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.460638] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.461536] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 648.461536] env[62066]: Removing descriptor: 18 [ 648.684825] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.697037] env[62066]: DEBUG nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.710751] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.710994] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.711191] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.711386] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.711536] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.711683] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.711885] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.712050] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.712217] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.712378] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.712588] env[62066]: DEBUG nova.virt.hardware [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.713497] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbb8a43-a239-4e23-853f-054fce95f27e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.721706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97783b12-65ef-4e52-8cc0-0d777729436a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.737552] env[62066]: ERROR nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Traceback (most recent call last): [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] yield resources [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.driver.spawn(context, instance, image_meta, [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] vm_ref = self.build_virtual_machine(instance, [ 648.737552] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] for vif in network_info: [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return self._sync_wrapper(fn, *args, **kwargs) [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.wait() [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self[:] = self._gt.wait() [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return self._exit_event.wait() [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.737938] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] current.throw(*self._exc) [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] result = function(*args, **kwargs) [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return func(*args, **kwargs) [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise e [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] nwinfo = self.network_api.allocate_for_instance( [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] created_port_ids = self._update_ports_for_instance( [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] with excutils.save_and_reraise_exception(): [ 648.738333] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.force_reraise() [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise self.value [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] updated_port = self._update_port( [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] _ensure_no_port_binding_failure(port) [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise exception.PortBindingFailed(port_id=port['id']) [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 648.738648] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] [ 648.738648] env[62066]: INFO nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Terminating instance [ 648.740971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquiring lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.741139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquired lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.741304] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.953989] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155725, 'name': ReconfigVM_Task, 'duration_secs': 0.272532} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.954291] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Reconfigured VM instance instance-0000001a to attach disk [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28/da7e4be3-29fc-462c-b8c7-ce4d80d19a28.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 648.954887] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7167965f-d03e-481f-bfca-05591b63473b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.960901] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 648.960901] env[62066]: value = "task-1155726" [ 648.960901] env[62066]: _type = "Task" [ 648.960901] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.968869] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155726, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.204640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.205194] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.207998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.648s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.263924] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.273459] env[62066]: DEBUG nova.compute.manager [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Received event network-changed-a8eed66c-0121-4e69-8d35-0afd4c422342 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.273654] env[62066]: DEBUG nova.compute.manager [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Refreshing instance network info cache due to event network-changed-a8eed66c-0121-4e69-8d35-0afd4c422342. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.273839] env[62066]: DEBUG oslo_concurrency.lockutils [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] Acquiring lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.356674] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.470381] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155726, 'name': Rename_Task, 'duration_secs': 0.13966} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.470669] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 649.470906] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7e7765d-1040-446a-9c51-f31b5994e40c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.476823] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Waiting for the task: (returnval){ [ 649.476823] env[62066]: value = "task-1155727" [ 649.476823] env[62066]: _type = "Task" [ 649.476823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.484444] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155727, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.711493] env[62066]: DEBUG nova.compute.utils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 649.717480] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.717665] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 649.758229] env[62066]: DEBUG nova.policy [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0bd92c6fd05488e9bf8f95b97077bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e781d81f23142d4859c188db11c59b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 649.860995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Releasing lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.860995] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.861418] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.861859] env[62066]: DEBUG oslo_concurrency.lockutils [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] Acquired lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.864018] env[62066]: DEBUG nova.network.neutron [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Refreshing network info cache for port a8eed66c-0121-4e69-8d35-0afd4c422342 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 649.864018] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b701eb3c-9b89-4c77-a371-09e8406325df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.877225] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ec6279-e329-4fa5-a3ec-dc4f24d16e9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.899146] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ada6c82c-c4fd-41a2-b30c-3aa505af6383 could not be found. [ 649.899700] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.900046] env[62066]: INFO nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Took 0.04 seconds to destroy the instance on the hypervisor. [ 649.900463] env[62066]: DEBUG oslo.service.loopingcall [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.900843] env[62066]: DEBUG nova.compute.manager [-] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.901077] env[62066]: DEBUG nova.network.neutron [-] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.920928] env[62066]: DEBUG nova.network.neutron [-] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.990129] env[62066]: DEBUG oslo_vmware.api [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Task: {'id': task-1155727, 'name': PowerOnVM_Task, 'duration_secs': 0.441539} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.990661] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 649.991000] env[62066]: DEBUG nova.compute.manager [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 649.991968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd212dd1-df24-4181-a017-40b9e1a6dad6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.064069] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Successfully created port: 7415a7f4-19d2-4e6e-81ba-d32423c579de {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.223203] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.269579] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance b4cb3d6d-d8ec-4632-8020-fae39b18143e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.269764] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2dc8d357-2864-410a-93ef-21ea4e9b530c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 650.269891] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance da7e4be3-29fc-462c-b8c7-ce4d80d19a28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270020] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 58528227-a9dc-4c12-b15b-fa191bc29046 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270137] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a406e14c-b281-4d1f-97d0-3219c6ffaa8d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270252] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d347bf40-941c-428f-85c3-18067a21d6c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270366] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a943544f-28d3-4e3f-8d95-0b1cc6f2c59d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270478] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance e0b6deda-1870-44f3-b39a-e971d592be9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270633] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ada6c82c-c4fd-41a2-b30c-3aa505af6383 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.270745] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3f0bf344-d408-471d-87e2-d3e0ef5c40fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 650.388663] env[62066]: DEBUG nova.network.neutron [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.424122] env[62066]: DEBUG nova.network.neutron [-] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.511659] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.525013] env[62066]: DEBUG nova.network.neutron [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.774328] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 9609ae16-202e-439b-a7e4-c1e7a0254a74 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.926490] env[62066]: INFO nova.compute.manager [-] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Took 1.03 seconds to deallocate network for instance. [ 650.931915] env[62066]: DEBUG nova.compute.claims [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.932105] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.026945] env[62066]: DEBUG oslo_concurrency.lockutils [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] Releasing lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.027233] env[62066]: DEBUG nova.compute.manager [req-49a9e446-a252-475f-bf9c-3a8c7b280108 req-169ccef5-2b4f-4b2d-84ee-f603b5ac0981 service nova] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Received event network-vif-deleted-a8eed66c-0121-4e69-8d35-0afd4c422342 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.236810] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.272437] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.272437] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.272437] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.272595] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.272595] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.272595] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.272595] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.272595] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.272765] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.273150] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.273490] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.275319] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661b2bb4-63b3-4a75-b198-ec4eca36f89a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.279858] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance e7b9c093-03a1-4cf9-bca5-41413aa4bcd8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.288889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4046b52a-8a24-4e8f-abd3-57aee83f9b94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.390137] env[62066]: DEBUG nova.compute.manager [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Received event network-changed-7415a7f4-19d2-4e6e-81ba-d32423c579de {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.390137] env[62066]: DEBUG nova.compute.manager [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Refreshing instance network info cache due to event network-changed-7415a7f4-19d2-4e6e-81ba-d32423c579de. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.390137] env[62066]: DEBUG oslo_concurrency.lockutils [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] Acquiring lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.390137] env[62066]: DEBUG oslo_concurrency.lockutils [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] Acquired lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.390137] env[62066]: DEBUG nova.network.neutron [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Refreshing network info cache for port 7415a7f4-19d2-4e6e-81ba-d32423c579de {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 651.478918] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 651.478918] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.478918] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.478918] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.478918] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.478918] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.478918] env[62066]: ERROR nova.compute.manager raise self.value [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.478918] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.478918] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.478918] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.479434] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.479434] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.479434] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 651.479434] env[62066]: ERROR nova.compute.manager [ 651.479434] env[62066]: Traceback (most recent call last): [ 651.479434] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.479434] env[62066]: listener.cb(fileno) [ 651.479434] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.479434] env[62066]: result = function(*args, **kwargs) [ 651.479434] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.479434] env[62066]: return func(*args, **kwargs) [ 651.479434] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.479434] env[62066]: raise e [ 651.479434] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.479434] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 651.479434] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.479434] env[62066]: created_port_ids = self._update_ports_for_instance( [ 651.479434] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.479434] env[62066]: with excutils.save_and_reraise_exception(): [ 651.479434] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.479434] env[62066]: self.force_reraise() [ 651.479434] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.479434] env[62066]: raise self.value [ 651.479434] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.479434] env[62066]: updated_port = self._update_port( [ 651.479434] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.479434] env[62066]: _ensure_no_port_binding_failure(port) [ 651.479434] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.479434] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.480190] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 651.480190] env[62066]: Removing descriptor: 18 [ 651.485056] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Traceback (most recent call last): [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] yield resources [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.driver.spawn(context, instance, image_meta, [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] vm_ref = self.build_virtual_machine(instance, [ 651.485056] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] for vif in network_info: [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self._sync_wrapper(fn, *args, **kwargs) [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.wait() [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self[:] = self._gt.wait() [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self._exit_event.wait() [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.485423] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] result = hub.switch() [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self.greenlet.switch() [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] result = function(*args, **kwargs) [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return func(*args, **kwargs) [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise e [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] nwinfo = self.network_api.allocate_for_instance( [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] created_port_ids = self._update_ports_for_instance( [ 651.485744] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] with excutils.save_and_reraise_exception(): [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.force_reraise() [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise self.value [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] updated_port = self._update_port( [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] _ensure_no_port_binding_failure(port) [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise exception.PortBindingFailed(port_id=port['id']) [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 651.486100] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] [ 651.486539] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Terminating instance [ 651.486539] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.755387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.755638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.755930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.756268] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.756420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.759398] env[62066]: INFO nova.compute.manager [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Terminating instance [ 651.761202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "refresh_cache-da7e4be3-29fc-462c-b8c7-ce4d80d19a28" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.761354] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquired lock "refresh_cache-da7e4be3-29fc-462c-b8c7-ce4d80d19a28" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.761519] env[62066]: DEBUG nova.network.neutron [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.784717] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d584bc1e-b5a3-4885-a6fb-efec8267d894 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.908774] env[62066]: DEBUG nova.network.neutron [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.990067] env[62066]: DEBUG nova.network.neutron [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.283251] env[62066]: DEBUG nova.network.neutron [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.287633] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d6837004-9b24-45f7-9ed7-48c4503859cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.331882] env[62066]: DEBUG nova.network.neutron [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.492443] env[62066]: DEBUG oslo_concurrency.lockutils [req-3100fff1-11f3-4d05-ae1e-078b80f4669b req-f3b861e8-685f-418b-ac28-20c7c271e12f service nova] Releasing lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.492912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.493136] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.774902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.775027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.790448] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ba33e43a-fbed-4a5a-9c05-a5a29082ebe9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.834658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Releasing lock "refresh_cache-da7e4be3-29fc-462c-b8c7-ce4d80d19a28" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.835159] env[62066]: DEBUG nova.compute.manager [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.835394] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.836654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb636fd-5bc4-452e-9f12-117bf8697691 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.845049] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 652.845328] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9aad222-088d-4a43-bce9-2c1cc7abac9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.851853] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 652.851853] env[62066]: value = "task-1155728" [ 652.851853] env[62066]: _type = "Task" [ 652.851853] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.861037] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.012595] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.102024] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.294020] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance c0ac4362-766f-48ba-aeb2-7fd976c1f47f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.362291] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155728, 'name': PowerOffVM_Task, 'duration_secs': 0.182063} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.362558] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 653.362725] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 653.363089] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3ef1895-c542-41af-aab4-604b67db65d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.387748] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 653.387950] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 653.388146] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Deleting the datastore file [datastore2] da7e4be3-29fc-462c-b8c7-ce4d80d19a28 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 653.388409] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac9f7960-fdd0-49f0-87ea-1cd7fd6698c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.394814] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for the task: (returnval){ [ 653.394814] env[62066]: value = "task-1155730" [ 653.394814] env[62066]: _type = "Task" [ 653.394814] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.402474] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155730, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.463709] env[62066]: DEBUG nova.compute.manager [req-447765c5-5a66-4e38-baa0-ad784e94e031 req-c8f50a85-a757-4d70-8dbb-356b3106e959 service nova] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Received event network-vif-deleted-7415a7f4-19d2-4e6e-81ba-d32423c579de {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.605206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.605802] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.606903] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.607201] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1b03d87-8fc7-4573-8851-7f76740a03e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.616858] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d1ec2d-6130-47ae-b725-7273647c97e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.640892] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f0bf344-d408-471d-87e2-d3e0ef5c40fd could not be found. [ 653.641017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.641293] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 653.641610] env[62066]: DEBUG oslo.service.loopingcall [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.641851] env[62066]: DEBUG nova.compute.manager [-] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.641947] env[62066]: DEBUG nova.network.neutron [-] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.659782] env[62066]: DEBUG nova.network.neutron [-] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.796670] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.904856] env[62066]: DEBUG oslo_vmware.api [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Task: {'id': task-1155730, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096424} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.905125] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 653.905383] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 653.905503] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.905675] env[62066]: INFO nova.compute.manager [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Took 1.07 seconds to destroy the instance on the hypervisor. [ 653.905904] env[62066]: DEBUG oslo.service.loopingcall [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.906098] env[62066]: DEBUG nova.compute.manager [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.906212] env[62066]: DEBUG nova.network.neutron [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.921315] env[62066]: DEBUG nova.network.neutron [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.162371] env[62066]: DEBUG nova.network.neutron [-] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.299955] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.423272] env[62066]: DEBUG nova.network.neutron [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.665249] env[62066]: INFO nova.compute.manager [-] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Took 1.02 seconds to deallocate network for instance. [ 654.667602] env[62066]: DEBUG nova.compute.claims [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.667824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.803400] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2f186d0a-91a0-4dc2-83bd-511099445af7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.926352] env[62066]: INFO nova.compute.manager [-] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Took 1.02 seconds to deallocate network for instance. [ 655.306251] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.433393] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.809075] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d50f0505-66f4-412f-9744-25e0ea96277a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.312330] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a8dd7483-0588-4f60-9504-20de799e69f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.815854] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 265906be-704e-4f0f-ae33-e89135e27268 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.318778] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.821889] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6a498481-b2ca-4813-87b7-2f09dfa107f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.324650] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.827670] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 641bf537-76d2-4c41-a6dc-b5b042327477 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.827950] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 658.828111] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 659.143906] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f9d5aa-b23e-4de5-b358-64f24534a5e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.151384] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c2f61b-bcbc-4d0f-a2a3-e216735c13bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.180258] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b55ccb2-7a5e-44f9-ac9c-2ce43833956a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.186658] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafed5ca-af96-4c04-89cf-b38e4a2529f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.198927] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.701907] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.207071] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 660.207357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.999s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.207645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.638s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.991019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c9b975-8caf-4000-841a-56516c3f5113 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.999028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76dc7078-e448-4040-94d2-6f2fd934e245 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.028236] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56242892-dc5d-4b64-a568-fc89d57932e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.035351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7b4e76-5bb2-44c5-a1ad-8e7d529bedc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.048199] env[62066]: DEBUG nova.compute.provider_tree [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.551090] env[62066]: DEBUG nova.scheduler.client.report [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.055948] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.848s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.056928] env[62066]: ERROR nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Traceback (most recent call last): [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.driver.spawn(context, instance, image_meta, [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] vm_ref = self.build_virtual_machine(instance, [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.056928] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] for vif in network_info: [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self._sync_wrapper(fn, *args, **kwargs) [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.wait() [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self[:] = self._gt.wait() [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self._exit_event.wait() [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] result = hub.switch() [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.057366] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return self.greenlet.switch() [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] result = function(*args, **kwargs) [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] return func(*args, **kwargs) [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise e [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] nwinfo = self.network_api.allocate_for_instance( [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] created_port_ids = self._update_ports_for_instance( [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] with excutils.save_and_reraise_exception(): [ 662.057679] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] self.force_reraise() [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise self.value [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] updated_port = self._update_port( [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] _ensure_no_port_binding_failure(port) [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] raise exception.PortBindingFailed(port_id=port['id']) [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] nova.exception.PortBindingFailed: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. [ 662.057995] env[62066]: ERROR nova.compute.manager [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] [ 662.058281] env[62066]: DEBUG nova.compute.utils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 662.058625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.164s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.060086] env[62066]: INFO nova.compute.claims [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.063051] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Build of instance 58528227-a9dc-4c12-b15b-fa191bc29046 was re-scheduled: Binding failed for port d0da24d6-830a-484a-af48-e42459fd3b21, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 662.063624] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 662.063719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquiring lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.063835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Acquired lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.064011] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.582646] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.653842] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.156662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Releasing lock "refresh_cache-58528227-a9dc-4c12-b15b-fa191bc29046" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.156946] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.157251] env[62066]: DEBUG nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.157471] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.176530] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.420023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d740fa1-e53b-46c4-9239-0c50777c98db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.427459] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c214cd-0478-4657-a2d6-f56c6835bace {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.458162] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5cae94-d2a3-40d7-a249-aa8e6f35889e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.465340] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6368c140-e949-4085-9530-98766eb50294 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.478234] env[62066]: DEBUG nova.compute.provider_tree [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.679822] env[62066]: DEBUG nova.network.neutron [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.981464] env[62066]: DEBUG nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.182872] env[62066]: INFO nova.compute.manager [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] [instance: 58528227-a9dc-4c12-b15b-fa191bc29046] Took 1.03 seconds to deallocate network for instance. [ 664.486495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.487008] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.489640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.232s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.994365] env[62066]: DEBUG nova.compute.utils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 664.998895] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.999086] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 665.047272] env[62066]: DEBUG nova.policy [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0bd92c6fd05488e9bf8f95b97077bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e781d81f23142d4859c188db11c59b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 665.211970] env[62066]: INFO nova.scheduler.client.report [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Deleted allocations for instance 58528227-a9dc-4c12-b15b-fa191bc29046 [ 665.328725] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Successfully created port: e7cf8d32-1b94-4da5-86a2-27b95c3d71d8 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.433917] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649cb96f-f5ac-4eac-a775-51f00e0ac00c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.443035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a76f29-4ded-40f2-ae90-a66ec0b27d30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.474365] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5436d0-fabf-47b4-b316-d4fe42cd3b99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.481589] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f8ce82-e328-43ec-9d1c-934c5c6df8c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.496377] env[62066]: DEBUG nova.compute.provider_tree [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.499384] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.722403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2cfd5b-71ef-4c3a-89ec-a2fec997902d tempest-ServerGroupTestJSON-1303192543 tempest-ServerGroupTestJSON-1303192543-project-member] Lock "58528227-a9dc-4c12-b15b-fa191bc29046" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.338s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.999810] env[62066]: DEBUG nova.scheduler.client.report [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.038656] env[62066]: DEBUG nova.compute.manager [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Received event network-changed-e7cf8d32-1b94-4da5-86a2-27b95c3d71d8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.038863] env[62066]: DEBUG nova.compute.manager [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Refreshing instance network info cache due to event network-changed-e7cf8d32-1b94-4da5-86a2-27b95c3d71d8. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 666.039088] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] Acquiring lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.039232] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] Acquired lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.039388] env[62066]: DEBUG nova.network.neutron [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Refreshing network info cache for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 666.192298] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 666.192298] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.192298] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.192298] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.192298] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.192298] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.192298] env[62066]: ERROR nova.compute.manager raise self.value [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.192298] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.192298] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.192298] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.192805] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.192805] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.192805] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 666.192805] env[62066]: ERROR nova.compute.manager [ 666.192805] env[62066]: Traceback (most recent call last): [ 666.192805] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.192805] env[62066]: listener.cb(fileno) [ 666.192805] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.192805] env[62066]: result = function(*args, **kwargs) [ 666.192805] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.192805] env[62066]: return func(*args, **kwargs) [ 666.192805] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.192805] env[62066]: raise e [ 666.192805] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.192805] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 666.192805] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.192805] env[62066]: created_port_ids = self._update_ports_for_instance( [ 666.192805] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.192805] env[62066]: with excutils.save_and_reraise_exception(): [ 666.192805] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.192805] env[62066]: self.force_reraise() [ 666.192805] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.192805] env[62066]: raise self.value [ 666.192805] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.192805] env[62066]: updated_port = self._update_port( [ 666.192805] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.192805] env[62066]: _ensure_no_port_binding_failure(port) [ 666.192805] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.192805] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.193648] env[62066]: nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 666.193648] env[62066]: Removing descriptor: 18 [ 666.225216] env[62066]: DEBUG nova.compute.manager [None req-8f33a439-9016-42da-bcd2-fe2a697691a5 tempest-ServersListShow296Test-8666260 tempest-ServersListShow296Test-8666260-project-member] [instance: ee5e1fd5-a946-4f00-ac06-8b3973873418] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.506967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.507673] env[62066]: ERROR nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Traceback (most recent call last): [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.driver.spawn(context, instance, image_meta, [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] vm_ref = self.build_virtual_machine(instance, [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.507673] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] for vif in network_info: [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self._sync_wrapper(fn, *args, **kwargs) [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.wait() [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self[:] = self._gt.wait() [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self._exit_event.wait() [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] result = hub.switch() [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.507960] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return self.greenlet.switch() [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] result = function(*args, **kwargs) [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] return func(*args, **kwargs) [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise e [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] nwinfo = self.network_api.allocate_for_instance( [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] created_port_ids = self._update_ports_for_instance( [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] with excutils.save_and_reraise_exception(): [ 666.508297] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] self.force_reraise() [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise self.value [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] updated_port = self._update_port( [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] _ensure_no_port_binding_failure(port) [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] raise exception.PortBindingFailed(port_id=port['id']) [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] nova.exception.PortBindingFailed: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. [ 666.508643] env[62066]: ERROR nova.compute.manager [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] [ 666.508917] env[62066]: DEBUG nova.compute.utils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 666.511187] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.513135] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.994s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.514734] env[62066]: INFO nova.compute.claims [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.517771] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Build of instance a406e14c-b281-4d1f-97d0-3219c6ffaa8d was re-scheduled: Binding failed for port a18da785-d71f-443d-997d-a3aa7f5c6797, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.518478] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.518478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.518579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.518731] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.544510] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.544982] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.545869] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.546466] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.546865] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.547167] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.547502] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.547678] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.547852] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.548075] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.548245] env[62066]: DEBUG nova.virt.hardware [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.551353] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3725af-7900-414d-a7c4-54e246cc23e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.560650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd6c749-7351-44c9-854b-c3aaad3fa52f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.577662] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Traceback (most recent call last): [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] yield resources [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.driver.spawn(context, instance, image_meta, [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] vm_ref = self.build_virtual_machine(instance, [ 666.577662] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] for vif in network_info: [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return self._sync_wrapper(fn, *args, **kwargs) [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.wait() [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self[:] = self._gt.wait() [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return self._exit_event.wait() [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 666.578104] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] current.throw(*self._exc) [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] result = function(*args, **kwargs) [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return func(*args, **kwargs) [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise e [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] nwinfo = self.network_api.allocate_for_instance( [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] created_port_ids = self._update_ports_for_instance( [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] with excutils.save_and_reraise_exception(): [ 666.578458] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.force_reraise() [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise self.value [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] updated_port = self._update_port( [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] _ensure_no_port_binding_failure(port) [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise exception.PortBindingFailed(port_id=port['id']) [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 666.578797] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] [ 666.578797] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Terminating instance [ 666.580120] env[62066]: DEBUG nova.network.neutron [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.582261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.668573] env[62066]: DEBUG nova.network.neutron [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.729831] env[62066]: DEBUG nova.compute.manager [None req-8f33a439-9016-42da-bcd2-fe2a697691a5 tempest-ServersListShow296Test-8666260 tempest-ServersListShow296Test-8666260-project-member] [instance: ee5e1fd5-a946-4f00-ac06-8b3973873418] Instance disappeared before build. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 667.038105] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.126130] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.171361] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14349c6-43b5-478f-9599-7790c7ea1abd req-b43c77ed-b167-48a8-b44f-85fdec5bdb90 service nova] Releasing lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.172059] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.172235] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 667.244276] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8f33a439-9016-42da-bcd2-fe2a697691a5 tempest-ServersListShow296Test-8666260 tempest-ServersListShow296Test-8666260-project-member] Lock "ee5e1fd5-a946-4f00-ac06-8b3973873418" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.315s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.628486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-a406e14c-b281-4d1f-97d0-3219c6ffaa8d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.628486] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.628662] env[62066]: DEBUG nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.628701] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.649998] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.690466] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.747199] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.922426] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.960735] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077da53c-17c5-492c-81c2-10bd0740703c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.969811] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12804419-9d9c-4104-a370-9db2f9918783 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.001477] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6128d0bb-b68a-4422-a110-e7f5c246d911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.008471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d920d9-9c62-4b8c-9b21-fb7428540df7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.023285] env[62066]: DEBUG nova.compute.provider_tree [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.098937] env[62066]: DEBUG nova.compute.manager [req-1469a0fc-6830-4945-8daa-3418d98f7fd7 req-3d0bf150-38f3-44e3-a7f4-ed7d283ada2a service nova] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Received event network-vif-deleted-e7cf8d32-1b94-4da5-86a2-27b95c3d71d8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.152948] env[62066]: DEBUG nova.network.neutron [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.271016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.425353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.425835] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 668.426267] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.426584] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f72dbe09-37c3-490b-9bdf-baff2e262f2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.435830] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268f9274-adac-459b-92ea-6d6e68d42daa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.458225] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9609ae16-202e-439b-a7e4-c1e7a0254a74 could not be found. [ 668.458428] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.458611] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Took 0.03 seconds to destroy the instance on the hypervisor. [ 668.458870] env[62066]: DEBUG oslo.service.loopingcall [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.459113] env[62066]: DEBUG nova.compute.manager [-] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.459207] env[62066]: DEBUG nova.network.neutron [-] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.478776] env[62066]: DEBUG nova.network.neutron [-] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.526205] env[62066]: DEBUG nova.scheduler.client.report [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.655828] env[62066]: INFO nova.compute.manager [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: a406e14c-b281-4d1f-97d0-3219c6ffaa8d] Took 1.03 seconds to deallocate network for instance. [ 668.981534] env[62066]: DEBUG nova.network.neutron [-] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.031256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.031850] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.034698] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.220s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.034922] env[62066]: DEBUG nova.objects.instance [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 669.485024] env[62066]: INFO nova.compute.manager [-] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Took 1.02 seconds to deallocate network for instance. [ 669.486642] env[62066]: DEBUG nova.compute.claims [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 669.486883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.542023] env[62066]: DEBUG nova.compute.utils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 669.544211] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.544394] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.626434] env[62066]: DEBUG nova.policy [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d7591cc2314ab3a606006c37942956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88cb350d909a48119d65c2f4232b4cb8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.694098] env[62066]: INFO nova.scheduler.client.report [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted allocations for instance a406e14c-b281-4d1f-97d0-3219c6ffaa8d [ 670.045044] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.049079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-12ea7862-d88a-42a6-9213-06a483121d23 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.055019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.558s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.055019] env[62066]: DEBUG nova.objects.instance [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 670.168610] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Successfully created port: a29f546d-f1f4-4728-b265-084cd4781a8d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.207344] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6fe3c2e2-b9d9-49db-9678-ddb28f6f608c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "a406e14c-b281-4d1f-97d0-3219c6ffaa8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.908s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.708373] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 671.059883] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.068212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c113350a-8ce3-4e2a-8c1b-38f3279b2e16 tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.068212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.024s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.093285] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.093285] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.093444] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.093617] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.093878] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.093939] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.094145] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.096395] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.096395] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.096395] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.096395] env[62066]: DEBUG nova.virt.hardware [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.096395] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3808f9e5-e760-4832-9c81-46e97dec8afa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.105668] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ce3d9e-4f0c-47de-9987-309ca11193b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.221801] env[62066]: ERROR nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 671.221801] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.221801] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.221801] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.221801] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.221801] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.221801] env[62066]: ERROR nova.compute.manager raise self.value [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.221801] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.221801] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.221801] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.222256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.222256] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.222256] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 671.222256] env[62066]: ERROR nova.compute.manager [ 671.222256] env[62066]: Traceback (most recent call last): [ 671.222256] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.222256] env[62066]: listener.cb(fileno) [ 671.222256] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.222256] env[62066]: result = function(*args, **kwargs) [ 671.222256] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.222256] env[62066]: return func(*args, **kwargs) [ 671.222256] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.222256] env[62066]: raise e [ 671.222256] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.222256] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 671.222256] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.222256] env[62066]: created_port_ids = self._update_ports_for_instance( [ 671.222256] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.222256] env[62066]: with excutils.save_and_reraise_exception(): [ 671.222256] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.222256] env[62066]: self.force_reraise() [ 671.222256] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.222256] env[62066]: raise self.value [ 671.222256] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.222256] env[62066]: updated_port = self._update_port( [ 671.222256] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.222256] env[62066]: _ensure_no_port_binding_failure(port) [ 671.222256] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.222256] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.223097] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 671.223097] env[62066]: Removing descriptor: 18 [ 671.223097] env[62066]: ERROR nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Traceback (most recent call last): [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] yield resources [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.driver.spawn(context, instance, image_meta, [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.223097] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] vm_ref = self.build_virtual_machine(instance, [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] for vif in network_info: [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self._sync_wrapper(fn, *args, **kwargs) [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.wait() [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self[:] = self._gt.wait() [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self._exit_event.wait() [ 671.223401] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] result = hub.switch() [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self.greenlet.switch() [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] result = function(*args, **kwargs) [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return func(*args, **kwargs) [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise e [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] nwinfo = self.network_api.allocate_for_instance( [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.223717] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] created_port_ids = self._update_ports_for_instance( [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] with excutils.save_and_reraise_exception(): [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.force_reraise() [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise self.value [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] updated_port = self._update_port( [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] _ensure_no_port_binding_failure(port) [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.224099] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise exception.PortBindingFailed(port_id=port['id']) [ 671.224402] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 671.224402] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] [ 671.224402] env[62066]: INFO nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Terminating instance [ 671.225425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquiring lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.225635] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquired lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.225927] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.235042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.273330] env[62066]: DEBUG nova.compute.manager [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Received event network-changed-a29f546d-f1f4-4728-b265-084cd4781a8d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.273330] env[62066]: DEBUG nova.compute.manager [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Refreshing instance network info cache due to event network-changed-a29f546d-f1f4-4728-b265-084cd4781a8d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 671.273330] env[62066]: DEBUG oslo_concurrency.lockutils [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] Acquiring lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.624876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.625121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.751417] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.811398] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.998020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2da025-0e23-4c8a-abca-5121e0a545d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.002510] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813466ca-4284-4a0f-88bd-1cf1aa3979d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.032186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5cfd78-722a-44e1-aa5b-9d7b12342ffa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.039532] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd510dd-302d-43e5-9789-bfd32e66e760 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.052745] env[62066]: DEBUG nova.compute.provider_tree [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.315039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Releasing lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.315039] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.315039] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.315303] env[62066]: DEBUG oslo_concurrency.lockutils [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] Acquired lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.315481] env[62066]: DEBUG nova.network.neutron [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Refreshing network info cache for port a29f546d-f1f4-4728-b265-084cd4781a8d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 672.316892] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-085f987c-0c17-4da7-9690-859218291b4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.326748] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d12f95-df0d-4af3-ad4b-dc647d9beb75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.350161] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7b9c093-03a1-4cf9-bca5-41413aa4bcd8 could not be found. [ 672.350161] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.350274] env[62066]: INFO nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.350879] env[62066]: DEBUG oslo.service.loopingcall [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.350879] env[62066]: DEBUG nova.compute.manager [-] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.350879] env[62066]: DEBUG nova.network.neutron [-] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.368964] env[62066]: DEBUG nova.network.neutron [-] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.470302] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "248f5281-eb46-4be1-8642-28813c4b2622" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.470480] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.557194] env[62066]: DEBUG nova.scheduler.client.report [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.836767] env[62066]: DEBUG nova.network.neutron [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.870908] env[62066]: DEBUG nova.network.neutron [-] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.882206] env[62066]: DEBUG nova.network.neutron [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.062224] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.062880] env[62066]: ERROR nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Traceback (most recent call last): [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.driver.spawn(context, instance, image_meta, [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] vm_ref = self.build_virtual_machine(instance, [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.062880] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] for vif in network_info: [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return self._sync_wrapper(fn, *args, **kwargs) [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.wait() [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self[:] = self._gt.wait() [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return self._exit_event.wait() [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] current.throw(*self._exc) [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.063281] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] result = function(*args, **kwargs) [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] return func(*args, **kwargs) [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise e [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] nwinfo = self.network_api.allocate_for_instance( [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] created_port_ids = self._update_ports_for_instance( [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] with excutils.save_and_reraise_exception(): [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] self.force_reraise() [ 673.063796] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise self.value [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] updated_port = self._update_port( [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] _ensure_no_port_binding_failure(port) [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] raise exception.PortBindingFailed(port_id=port['id']) [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] nova.exception.PortBindingFailed: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. [ 673.064170] env[62066]: ERROR nova.compute.manager [instance: d347bf40-941c-428f-85c3-18067a21d6c9] [ 673.064170] env[62066]: DEBUG nova.compute.utils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 673.064944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.259s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.066364] env[62066]: INFO nova.compute.claims [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.069093] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Build of instance d347bf40-941c-428f-85c3-18067a21d6c9 was re-scheduled: Binding failed for port 3c6574e9-8721-409b-846a-39cf31c97460, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.069514] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.069751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquiring lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.069895] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Acquired lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.070061] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.293654] env[62066]: DEBUG nova.compute.manager [req-cc078855-0dbe-47fd-8637-b3bd5633f2a0 req-d3f01c37-350a-48fc-9b1e-8e51abb4166e service nova] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Received event network-vif-deleted-a29f546d-f1f4-4728-b265-084cd4781a8d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.373015] env[62066]: INFO nova.compute.manager [-] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Took 1.02 seconds to deallocate network for instance. [ 673.375264] env[62066]: DEBUG nova.compute.claims [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.375579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.384319] env[62066]: DEBUG oslo_concurrency.lockutils [req-d47ae250-826b-45eb-89e2-ab6e02e7c8c2 req-69d2e2f7-c84e-451a-89b1-570d8a082494 service nova] Releasing lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.588107] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.676617] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.179528] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Releasing lock "refresh_cache-d347bf40-941c-428f-85c3-18067a21d6c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.179799] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.179966] env[62066]: DEBUG nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.180145] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.194454] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.382379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409eca24-9ff9-4bbb-853b-d459cd75ffa1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.390043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf85287-93af-4d59-a0ec-03fbfacc6ce3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.420514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2414818-507e-4ee1-983d-fefccc1d060f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.427807] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061decec-1601-42a2-8bb4-bbadef19004a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.440915] env[62066]: DEBUG nova.compute.provider_tree [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.698466] env[62066]: DEBUG nova.network.neutron [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.944415] env[62066]: DEBUG nova.scheduler.client.report [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.201373] env[62066]: INFO nova.compute.manager [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] [instance: d347bf40-941c-428f-85c3-18067a21d6c9] Took 1.02 seconds to deallocate network for instance. [ 675.449263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.449807] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.452334] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.343s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.452493] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.454470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.945s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.472969] env[62066]: INFO nova.scheduler.client.report [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted allocations for instance 2dc8d357-2864-410a-93ef-21ea4e9b530c [ 675.959782] env[62066]: DEBUG nova.compute.utils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 675.964309] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.964450] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.983162] env[62066]: DEBUG oslo_concurrency.lockutils [None req-287fc675-0a54-45e0-ac32-45617030684d tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "2dc8d357-2864-410a-93ef-21ea4e9b530c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.598s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.004574] env[62066]: DEBUG nova.policy [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75c4c346db5c4f91a4be01a25b1ceec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad252c21f8f54840ac296964a89bd0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.243267] env[62066]: INFO nova.scheduler.client.report [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Deleted allocations for instance d347bf40-941c-428f-85c3-18067a21d6c9 [ 676.284152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.284325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.284575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.285015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.285015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.287263] env[62066]: INFO nova.compute.manager [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Terminating instance [ 676.289512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "refresh_cache-b4cb3d6d-d8ec-4632-8020-fae39b18143e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.289512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquired lock "refresh_cache-b4cb3d6d-d8ec-4632-8020-fae39b18143e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.289512] env[62066]: DEBUG nova.network.neutron [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.309957] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Successfully created port: 291a9f0b-2e9f-488a-8cc2-c31edc469a7a {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.313783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8462935-6852-4566-8028-08174f4bd422 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.323221] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae337ce3-6743-4dd6-abd7-7f9733542ae3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.354281] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60902ed-ec8e-4c8e-9c58-c030bd0e0b98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.361931] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d245c75-a7fb-44ff-a72b-23700cba9cf1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.375193] env[62066]: DEBUG nova.compute.provider_tree [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.465157] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.755159] env[62066]: DEBUG oslo_concurrency.lockutils [None req-167d3fe4-d0e8-4b5b-b196-2d3ec2202c3e tempest-ServersTestManualDisk-2024562636 tempest-ServersTestManualDisk-2024562636-project-member] Lock "d347bf40-941c-428f-85c3-18067a21d6c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.432s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.811191] env[62066]: DEBUG nova.network.neutron [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.878844] env[62066]: DEBUG nova.scheduler.client.report [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.889126] env[62066]: DEBUG nova.network.neutron [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.143674] env[62066]: DEBUG nova.compute.manager [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Received event network-changed-291a9f0b-2e9f-488a-8cc2-c31edc469a7a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.143873] env[62066]: DEBUG nova.compute.manager [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Refreshing instance network info cache due to event network-changed-291a9f0b-2e9f-488a-8cc2-c31edc469a7a. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 677.144562] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] Acquiring lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.144755] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] Acquired lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.144921] env[62066]: DEBUG nova.network.neutron [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Refreshing network info cache for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.201875] env[62066]: ERROR nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 677.201875] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.201875] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.201875] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.201875] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.201875] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.201875] env[62066]: ERROR nova.compute.manager raise self.value [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.201875] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.201875] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.201875] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.202335] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.202335] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.202335] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 677.202335] env[62066]: ERROR nova.compute.manager [ 677.202335] env[62066]: Traceback (most recent call last): [ 677.202335] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.202335] env[62066]: listener.cb(fileno) [ 677.202335] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.202335] env[62066]: result = function(*args, **kwargs) [ 677.202335] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.202335] env[62066]: return func(*args, **kwargs) [ 677.202335] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.202335] env[62066]: raise e [ 677.202335] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.202335] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 677.202335] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.202335] env[62066]: created_port_ids = self._update_ports_for_instance( [ 677.202335] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.202335] env[62066]: with excutils.save_and_reraise_exception(): [ 677.202335] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.202335] env[62066]: self.force_reraise() [ 677.202335] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.202335] env[62066]: raise self.value [ 677.202335] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.202335] env[62066]: updated_port = self._update_port( [ 677.202335] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.202335] env[62066]: _ensure_no_port_binding_failure(port) [ 677.202335] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.202335] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.203033] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 677.203033] env[62066]: Removing descriptor: 18 [ 677.257961] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.386022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.386485] env[62066]: ERROR nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Traceback (most recent call last): [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.driver.spawn(context, instance, image_meta, [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] vm_ref = self.build_virtual_machine(instance, [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.386485] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] for vif in network_info: [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return self._sync_wrapper(fn, *args, **kwargs) [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.wait() [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self[:] = self._gt.wait() [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return self._exit_event.wait() [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] current.throw(*self._exc) [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.386855] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] result = function(*args, **kwargs) [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] return func(*args, **kwargs) [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise e [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] nwinfo = self.network_api.allocate_for_instance( [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] created_port_ids = self._update_ports_for_instance( [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] with excutils.save_and_reraise_exception(): [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] self.force_reraise() [ 677.387367] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise self.value [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] updated_port = self._update_port( [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] _ensure_no_port_binding_failure(port) [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] raise exception.PortBindingFailed(port_id=port['id']) [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] nova.exception.PortBindingFailed: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. [ 677.388285] env[62066]: ERROR nova.compute.manager [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] [ 677.388285] env[62066]: DEBUG nova.compute.utils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 677.388631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.391443] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Build of instance a943544f-28d3-4e3f-8d95-0b1cc6f2c59d was re-scheduled: Binding failed for port dd3173d0-dd91-44a1-a785-43c57a362fbf, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.391859] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.392097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.392244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.392590] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.394778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Releasing lock "refresh_cache-b4cb3d6d-d8ec-4632-8020-fae39b18143e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.395662] env[62066]: DEBUG nova.compute.manager [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.395662] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.396205] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061e57a0-3662-45cc-9785-b30390072d0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.404825] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 677.405098] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa61bee7-b754-4b47-998d-b6db02a1c5ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.411778] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 677.411778] env[62066]: value = "task-1155731" [ 677.411778] env[62066]: _type = "Task" [ 677.411778] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.419351] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.477740] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.509144] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.509144] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.509144] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.509350] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.509350] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.509350] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.509350] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.509350] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.509486] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.509486] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.509610] env[62066]: DEBUG nova.virt.hardware [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.511768] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece487c2-8ecc-415d-add2-790cae594aad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.518460] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7867c49f-d24c-488c-b004-fffdc0fbe2d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.532829] env[62066]: ERROR nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Traceback (most recent call last): [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] yield resources [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.driver.spawn(context, instance, image_meta, [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] vm_ref = self.build_virtual_machine(instance, [ 677.532829] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] for vif in network_info: [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return self._sync_wrapper(fn, *args, **kwargs) [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.wait() [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self[:] = self._gt.wait() [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return self._exit_event.wait() [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.533212] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] current.throw(*self._exc) [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] result = function(*args, **kwargs) [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return func(*args, **kwargs) [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise e [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] nwinfo = self.network_api.allocate_for_instance( [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] created_port_ids = self._update_ports_for_instance( [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] with excutils.save_and_reraise_exception(): [ 677.533540] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.force_reraise() [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise self.value [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] updated_port = self._update_port( [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] _ensure_no_port_binding_failure(port) [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise exception.PortBindingFailed(port_id=port['id']) [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 677.533874] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] [ 677.533874] env[62066]: INFO nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Terminating instance [ 677.536077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquiring lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.672061] env[62066]: DEBUG nova.network.neutron [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.782049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.887809] env[62066]: DEBUG nova.network.neutron [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.916948] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.925402] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155731, 'name': PowerOffVM_Task, 'duration_secs': 0.114389} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.925966] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 677.925966] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 677.926090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f76bddf9-279b-4a49-989e-626d5119dbfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.953252] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 677.953731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 677.953731] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleting the datastore file [datastore2] b4cb3d6d-d8ec-4632-8020-fae39b18143e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 677.953880] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f70d1eaf-7e33-408d-b253-91ff261c97ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.961631] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for the task: (returnval){ [ 677.961631] env[62066]: value = "task-1155733" [ 677.961631] env[62066]: _type = "Task" [ 677.961631] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.969498] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.998272] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.296842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae13f50-7405-4dc0-ba30-a85105271444 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.304752] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba17e6b2-2082-4bca-8024-eab87aec06be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.334476] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be453f5-a767-472f-a994-2ec29135af91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.341907] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c468f33c-ebd6-48cb-91a9-80ab0cfdd108 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.355024] env[62066]: DEBUG nova.compute.provider_tree [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.392522] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e355b74-4bf7-4afb-915f-28e7a5d4141c req-c444d9e3-6e27-458c-b95a-a63e96391d9a service nova] Releasing lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.393382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquired lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.393382] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.471714] env[62066]: DEBUG oslo_vmware.api [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Task: {'id': task-1155733, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110163} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.472213] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 678.472591] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 678.472906] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.473233] env[62066]: INFO nova.compute.manager [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 678.473599] env[62066]: DEBUG oslo.service.loopingcall [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.473930] env[62066]: DEBUG nova.compute.manager [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.475435] env[62066]: DEBUG nova.network.neutron [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.491657] env[62066]: DEBUG nova.network.neutron [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.501476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "refresh_cache-a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.501476] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.501476] env[62066]: DEBUG nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.501476] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.515034] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.857501] env[62066]: DEBUG nova.scheduler.client.report [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.918500] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.992808] env[62066]: DEBUG nova.network.neutron [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.011140] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.018371] env[62066]: DEBUG nova.network.neutron [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.168479] env[62066]: DEBUG nova.compute.manager [req-a6ce7c7b-fc58-49d8-a602-2b66697e10df req-0afe2719-2960-4de1-b8bd-9d693734882c service nova] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Received event network-vif-deleted-291a9f0b-2e9f-488a-8cc2-c31edc469a7a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.362891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.363539] env[62066]: ERROR nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Traceback (most recent call last): [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.driver.spawn(context, instance, image_meta, [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] vm_ref = self.build_virtual_machine(instance, [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.363539] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] for vif in network_info: [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return self._sync_wrapper(fn, *args, **kwargs) [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.wait() [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self[:] = self._gt.wait() [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return self._exit_event.wait() [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] current.throw(*self._exc) [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.363898] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] result = function(*args, **kwargs) [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] return func(*args, **kwargs) [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise e [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] nwinfo = self.network_api.allocate_for_instance( [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] created_port_ids = self._update_ports_for_instance( [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] with excutils.save_and_reraise_exception(): [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] self.force_reraise() [ 679.364277] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise self.value [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] updated_port = self._update_port( [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] _ensure_no_port_binding_failure(port) [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] raise exception.PortBindingFailed(port_id=port['id']) [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] nova.exception.PortBindingFailed: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. [ 679.364638] env[62066]: ERROR nova.compute.manager [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] [ 679.364638] env[62066]: DEBUG nova.compute.utils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 679.365917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.854s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.366111] env[62066]: DEBUG nova.objects.instance [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 679.369670] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Build of instance e0b6deda-1870-44f3-b39a-e971d592be9e was re-scheduled: Binding failed for port cb97e1cb-5857-4b19-893b-5a360521e01c, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.370126] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.370402] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquiring lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.370622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Acquired lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.370739] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.495340] env[62066]: INFO nova.compute.manager [-] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Took 1.02 seconds to deallocate network for instance. [ 679.514120] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Releasing lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.514552] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.514849] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.515032] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f53942df-0673-45ed-8e57-797ca79fb933 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.526063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460f0760-5c24-4c53-ae60-12565e4815fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.541910] env[62066]: INFO nova.compute.manager [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a943544f-28d3-4e3f-8d95-0b1cc6f2c59d] Took 1.04 seconds to deallocate network for instance. [ 679.557950] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d584bc1e-b5a3-4885-a6fb-efec8267d894 could not be found. [ 679.558323] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.558506] env[62066]: INFO nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Took 0.04 seconds to destroy the instance on the hypervisor. [ 679.558776] env[62066]: DEBUG oslo.service.loopingcall [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.559009] env[62066]: DEBUG nova.compute.manager [-] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.559109] env[62066]: DEBUG nova.network.neutron [-] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.586157] env[62066]: DEBUG nova.network.neutron [-] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.891850] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.941491] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.001762] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.088376] env[62066]: DEBUG nova.network.neutron [-] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.378674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf5ebcac-8363-4557-8cc9-b6c9e0d9dab6 tempest-ServersAdmin275Test-615861432 tempest-ServersAdmin275Test-615861432-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.379789] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.448s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.445965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Releasing lock "refresh_cache-e0b6deda-1870-44f3-b39a-e971d592be9e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.445965] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.445965] env[62066]: DEBUG nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.445965] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.460235] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.575341] env[62066]: INFO nova.scheduler.client.report [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted allocations for instance a943544f-28d3-4e3f-8d95-0b1cc6f2c59d [ 680.591101] env[62066]: INFO nova.compute.manager [-] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Took 1.03 seconds to deallocate network for instance. [ 680.593498] env[62066]: DEBUG nova.compute.claims [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.593498] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.962633] env[62066]: DEBUG nova.network.neutron [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.087168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcfa0694-7a50-4f58-bfa7-c82e97ee0a68 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a943544f-28d3-4e3f-8d95-0b1cc6f2c59d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.770s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.196718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae71258-9594-409d-bdc7-dac78e1307b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.204600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54dfb47-167d-4994-b954-5b9d364d8dbe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.234650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42aaf93-4710-42a3-867a-2b26568b1271 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.242200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f62d974-0817-4444-bd92-cc244c9daa15 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.255711] env[62066]: DEBUG nova.compute.provider_tree [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.468385] env[62066]: INFO nova.compute.manager [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] [instance: e0b6deda-1870-44f3-b39a-e971d592be9e] Took 1.02 seconds to deallocate network for instance. [ 681.593188] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.762199] env[62066]: DEBUG nova.scheduler.client.report [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.120922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.265735] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.266671] env[62066]: ERROR nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Traceback (most recent call last): [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.driver.spawn(context, instance, image_meta, [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] vm_ref = self.build_virtual_machine(instance, [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.266671] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] for vif in network_info: [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return self._sync_wrapper(fn, *args, **kwargs) [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.wait() [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self[:] = self._gt.wait() [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return self._exit_event.wait() [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] current.throw(*self._exc) [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.266978] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] result = function(*args, **kwargs) [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] return func(*args, **kwargs) [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise e [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] nwinfo = self.network_api.allocate_for_instance( [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] created_port_ids = self._update_ports_for_instance( [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] with excutils.save_and_reraise_exception(): [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] self.force_reraise() [ 682.267313] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise self.value [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] updated_port = self._update_port( [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] _ensure_no_port_binding_failure(port) [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] raise exception.PortBindingFailed(port_id=port['id']) [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] nova.exception.PortBindingFailed: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. [ 682.267637] env[62066]: ERROR nova.compute.manager [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] [ 682.267637] env[62066]: DEBUG nova.compute.utils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 682.268969] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.601s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.273341] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Build of instance ada6c82c-c4fd-41a2-b30c-3aa505af6383 was re-scheduled: Binding failed for port a8eed66c-0121-4e69-8d35-0afd4c422342, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.275100] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.275461] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquiring lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.276282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Acquired lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.276467] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.504152] env[62066]: INFO nova.scheduler.client.report [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Deleted allocations for instance e0b6deda-1870-44f3-b39a-e971d592be9e [ 682.801617] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.001535] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.018955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2011481-ed06-4ade-b5da-b253413d4430 tempest-ServerExternalEventsTest-1230269119 tempest-ServerExternalEventsTest-1230269119-project-member] Lock "e0b6deda-1870-44f3-b39a-e971d592be9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.426s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.200695] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37366c1c-c410-4d5e-a3bd-18c066d3a966 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.209215] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474540f4-548f-4488-9300-9a90b7c6c56d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.241316] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5c9836-7be8-4b29-b4be-01ce9eb5b307 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.249238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd09f361-1660-4b1c-bfcc-ee38f28af4fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.266020] env[62066]: DEBUG nova.compute.provider_tree [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.504772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Releasing lock "refresh_cache-ada6c82c-c4fd-41a2-b30c-3aa505af6383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.505086] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.505231] env[62066]: DEBUG nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.505396] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.523078] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.529946] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.768534] env[62066]: DEBUG nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.036557] env[62066]: DEBUG nova.network.neutron [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.059026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.278540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.280672] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Traceback (most recent call last): [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.driver.spawn(context, instance, image_meta, [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] vm_ref = self.build_virtual_machine(instance, [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.280672] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] for vif in network_info: [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self._sync_wrapper(fn, *args, **kwargs) [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.wait() [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self[:] = self._gt.wait() [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self._exit_event.wait() [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] result = hub.switch() [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.281557] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return self.greenlet.switch() [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] result = function(*args, **kwargs) [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] return func(*args, **kwargs) [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise e [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] nwinfo = self.network_api.allocate_for_instance( [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] created_port_ids = self._update_ports_for_instance( [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] with excutils.save_and_reraise_exception(): [ 684.282534] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] self.force_reraise() [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise self.value [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] updated_port = self._update_port( [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] _ensure_no_port_binding_failure(port) [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] raise exception.PortBindingFailed(port_id=port['id']) [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] nova.exception.PortBindingFailed: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. [ 684.283193] env[62066]: ERROR nova.compute.manager [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] [ 684.285761] env[62066]: DEBUG nova.compute.utils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 684.287998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.855s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.289627] env[62066]: DEBUG nova.objects.instance [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lazy-loading 'resources' on Instance uuid da7e4be3-29fc-462c-b8c7-ce4d80d19a28 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.294510] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Build of instance 3f0bf344-d408-471d-87e2-d3e0ef5c40fd was re-scheduled: Binding failed for port 7415a7f4-19d2-4e6e-81ba-d32423c579de, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.294510] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.294510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.294510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.294937] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.372552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.372837] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.547903] env[62066]: INFO nova.compute.manager [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] [instance: ada6c82c-c4fd-41a2-b30c-3aa505af6383] Took 1.04 seconds to deallocate network for instance. [ 684.818791] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.913762] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.045065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "5db50c22-048b-4cce-962a-3df1262f6e4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.045314] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.137647] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a804f633-a2c5-4913-8cea-13cfef705d31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.145738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b365e7-6873-4e2b-bd02-a1b7e1b90538 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.176852] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1938ce3f-a991-4011-895e-39a549c83f3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.184113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a6d5ec-2ba4-458e-b453-86fe8d97c832 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.197715] env[62066]: DEBUG nova.compute.provider_tree [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.418046] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-3f0bf344-d408-471d-87e2-d3e0ef5c40fd" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.418046] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.418203] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.418249] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.433510] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.582315] env[62066]: INFO nova.scheduler.client.report [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Deleted allocations for instance ada6c82c-c4fd-41a2-b30c-3aa505af6383 [ 685.702070] env[62066]: DEBUG nova.scheduler.client.report [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.937154] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.090288] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2d4634a-de30-4381-840b-ece008f83d1a tempest-FloatingIPsAssociationTestJSON-2120219227 tempest-FloatingIPsAssociationTestJSON-2120219227-project-member] Lock "ada6c82c-c4fd-41a2-b30c-3aa505af6383" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.390s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.206583] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.209755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.211356] env[62066]: INFO nova.compute.claims [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.225909] env[62066]: INFO nova.scheduler.client.report [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Deleted allocations for instance da7e4be3-29fc-462c-b8c7-ce4d80d19a28 [ 686.440057] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 3f0bf344-d408-471d-87e2-d3e0ef5c40fd] Took 1.02 seconds to deallocate network for instance. [ 686.594307] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.732617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0f45f57c-fea9-41a9-9c52-cd2d5b7e1d3d tempest-ServersAdmin275Test-627033303 tempest-ServersAdmin275Test-627033303-project-member] Lock "da7e4be3-29fc-462c-b8c7-ce4d80d19a28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.977s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.123129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.477651] env[62066]: INFO nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted allocations for instance 3f0bf344-d408-471d-87e2-d3e0ef5c40fd [ 687.554112] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1bbad1-426f-4ac9-8903-850d6491aa1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.562218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b50a13-6e4c-4803-a555-a12a743884b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.600262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c6b2c3-3a27-4c9d-90a1-eb4d49f70e69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.607900] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f8cd6a-26b7-4080-9046-16bc5a6bf067 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.622866] env[62066]: DEBUG nova.compute.provider_tree [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.990055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "3f0bf344-d408-471d-87e2-d3e0ef5c40fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.382s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.126586] env[62066]: DEBUG nova.scheduler.client.report [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.497528] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.633171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.633703] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 688.637180] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.150s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.023278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.142187] env[62066]: DEBUG nova.compute.utils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 689.147701] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 689.481771] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590bd002-c019-4c82-92ad-607d9db3a110 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.489803] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1fbca3-bfc8-472f-aa19-fb15567e926a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.521607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da3a995-cc2d-4073-b1cc-ea85a2a3a60d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.528975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e598310e-7c6e-4feb-b6d8-d4d4be842125 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.543876] env[62066]: DEBUG nova.compute.provider_tree [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.648071] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.047078] env[62066]: DEBUG nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.552093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.552855] env[62066]: ERROR nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Traceback (most recent call last): [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.driver.spawn(context, instance, image_meta, [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] vm_ref = self.build_virtual_machine(instance, [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.552855] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] for vif in network_info: [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return self._sync_wrapper(fn, *args, **kwargs) [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.wait() [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self[:] = self._gt.wait() [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return self._exit_event.wait() [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] current.throw(*self._exc) [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.553267] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] result = function(*args, **kwargs) [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] return func(*args, **kwargs) [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise e [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] nwinfo = self.network_api.allocate_for_instance( [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] created_port_ids = self._update_ports_for_instance( [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] with excutils.save_and_reraise_exception(): [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] self.force_reraise() [ 690.553635] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise self.value [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] updated_port = self._update_port( [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] _ensure_no_port_binding_failure(port) [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] raise exception.PortBindingFailed(port_id=port['id']) [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] nova.exception.PortBindingFailed: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. [ 690.554032] env[62066]: ERROR nova.compute.manager [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] [ 690.554032] env[62066]: DEBUG nova.compute.utils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 690.554918] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.320s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.556356] env[62066]: INFO nova.compute.claims [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.562222] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Build of instance 9609ae16-202e-439b-a7e4-c1e7a0254a74 was re-scheduled: Binding failed for port e7cf8d32-1b94-4da5-86a2-27b95c3d71d8, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.562222] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.562222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.562222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.562472] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.660034] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 690.691235] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 690.691500] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 690.692564] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 690.692564] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 690.692564] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 690.692564] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 690.692564] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 690.692916] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 690.694217] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 690.694217] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 690.694217] env[62066]: DEBUG nova.virt.hardware [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.694968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcc3d78-f9e1-439c-a68a-2b369bac1296 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.707490] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3378f185-6714-4d73-86fb-b1e2159008cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.723070] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.729721] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Creating folder: Project (9cec5d77e27f4d9f80a258c772cfddf5). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.729721] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17c50c25-a00e-463b-919c-e585ce538156 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.741425] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Created folder: Project (9cec5d77e27f4d9f80a258c772cfddf5) in parent group-v251573. [ 690.741836] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Creating folder: Instances. Parent ref: group-v251591. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.742042] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4b3e8c6-24d2-45fb-b2cb-bc0c8c4ff7a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.749872] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Created folder: Instances in parent group-v251591. [ 690.750234] env[62066]: DEBUG oslo.service.loopingcall [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.754240] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 690.754464] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0568cbd-2786-4b4a-a4fa-1d7df32fe1c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.770680] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.770680] env[62066]: value = "task-1155736" [ 690.770680] env[62066]: _type = "Task" [ 690.770680] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.778115] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155736, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.092302] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.223653] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.282101] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155736, 'name': CreateVM_Task, 'duration_secs': 0.245889} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.282101] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 691.286851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.287182] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.287597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 691.288270] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d74ec16-2179-4577-bef4-4e1e710c47e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.293425] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 691.293425] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d7c563-9812-e51c-dcc9-b50cce3a1d28" [ 691.293425] env[62066]: _type = "Task" [ 691.293425] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.309024] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d7c563-9812-e51c-dcc9-b50cce3a1d28, 'name': SearchDatastore_Task, 'duration_secs': 0.00887} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.309024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.309024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.309024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.309218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.309218] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.309218] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34ca27d8-98ea-4827-aa3f-93b31b01d344 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.316074] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.316428] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 691.317721] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b1b001b-69d8-4d32-985d-31dc94f729fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.323656] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 691.323656] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ae19c9-26e9-1ac4-aa66-7bbf87d9b3ba" [ 691.323656] env[62066]: _type = "Task" [ 691.323656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.336913] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ae19c9-26e9-1ac4-aa66-7bbf87d9b3ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008061} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.337676] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f5c229d-d481-48f9-acee-f79a05cfe07d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.342511] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 691.342511] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cea9be-141a-811a-98ff-8762f566cb8e" [ 691.342511] env[62066]: _type = "Task" [ 691.342511] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.349795] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cea9be-141a-811a-98ff-8762f566cb8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.725632] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-9609ae16-202e-439b-a7e4-c1e7a0254a74" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.725885] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.726066] env[62066]: DEBUG nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.726203] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.748903] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.854012] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cea9be-141a-811a-98ff-8762f566cb8e, 'name': SearchDatastore_Task, 'duration_secs': 0.007843} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.854326] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.854586] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 691.854851] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abeb3998-8328-4b12-b6e6-bcb08af4b765 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.862146] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 691.862146] env[62066]: value = "task-1155737" [ 691.862146] env[62066]: _type = "Task" [ 691.862146] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.872719] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.941194] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247fee88-cfda-42c0-a62e-83d1608ea199 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.949280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3d7436-200b-4911-98c6-b83177438dd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.991443] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713ba21a-4c37-46db-9dcb-d9d72395641d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.001028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c73432f-9c2b-4528-8983-eb2b24935d2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.026789] env[62066]: DEBUG nova.compute.provider_tree [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.040409] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "1c694323-616c-43b1-bd49-bba707e0788f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.040600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.252694] env[62066]: DEBUG nova.network.neutron [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.371790] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.372525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 692.373018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.373291] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56ab2c2b-60a8-4c3b-8cc2-accd58409492 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.380516] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 692.380516] env[62066]: value = "task-1155738" [ 692.380516] env[62066]: _type = "Task" [ 692.380516] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.389779] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.530027] env[62066]: DEBUG nova.scheduler.client.report [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.608774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "3d001331-7a41-4385-a622-54ce481509a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.609844] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.754952] env[62066]: INFO nova.compute.manager [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 9609ae16-202e-439b-a7e4-c1e7a0254a74] Took 1.03 seconds to deallocate network for instance. [ 692.890634] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06236} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.890913] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.891794] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ca35cc-05ed-4cf4-9cad-cbc36288a461 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.911421] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.911713] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41b976c6-3e08-422f-a46a-b845d2b32f2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.932321] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 692.932321] env[62066]: value = "task-1155739" [ 692.932321] env[62066]: _type = "Task" [ 692.932321] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.940302] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.035083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.035636] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.038522] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.663s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.443806] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155739, 'name': ReconfigVM_Task, 'duration_secs': 0.264952} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.444126] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Reconfigured VM instance instance-00000026 to attach disk [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.444701] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6add5a7a-ce9f-44d5-ab57-714d9dff12fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.453376] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 693.453376] env[62066]: value = "task-1155740" [ 693.453376] env[62066]: _type = "Task" [ 693.453376] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.462256] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155740, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.543387] env[62066]: DEBUG nova.compute.utils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 693.548891] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.550118] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.591637] env[62066]: DEBUG nova.policy [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd26bafea7df4498da78677525d727ae4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a529a98d37ec468da1688b18be898029', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.789897] env[62066]: INFO nova.scheduler.client.report [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted allocations for instance 9609ae16-202e-439b-a7e4-c1e7a0254a74 [ 693.909121] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed23d95-148b-469e-a7ab-3b548375ad93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.918383] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06053e32-513a-471c-9bcd-4ef5bbc5049b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.921732] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Successfully created port: eee10f4a-57b6-40ad-8049-87544bda2f33 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.952156] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b0eb89-9bbc-403c-b263-726b7fcc606e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.965009] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bf5428-1584-4339-b3a1-2487a5180ca6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.968844] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155740, 'name': Rename_Task, 'duration_secs': 0.132469} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.969123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 693.969656] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30c9f875-889c-457e-9407-db9208a001f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.979967] env[62066]: DEBUG nova.compute.provider_tree [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.986720] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 693.986720] env[62066]: value = "task-1155741" [ 693.986720] env[62066]: _type = "Task" [ 693.986720] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.993414] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.050825] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.302662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-352abffc-fe91-4145-89c3-2bb43c5559c0 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "9609ae16-202e-439b-a7e4-c1e7a0254a74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.666s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.484440] env[62066]: DEBUG nova.scheduler.client.report [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.498032] env[62066]: DEBUG oslo_vmware.api [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155741, 'name': PowerOnVM_Task, 'duration_secs': 0.420993} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.498209] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 694.500035] env[62066]: INFO nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Took 3.84 seconds to spawn the instance on the hypervisor. [ 694.500035] env[62066]: DEBUG nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.500035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9743be28-f08c-4cd3-9b7a-00116fffd9f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.559614] env[62066]: INFO nova.virt.block_device [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Booting with volume 9afac259-c57e-46be-813d-343c3489eb54 at /dev/sda [ 694.607762] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91815da9-d27d-45b1-b63a-01d6cf73e8fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.617613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637a2d5b-8c2f-4103-9b55-d062469a316c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.647805] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc2d3f2a-a569-43b0-b834-1da0b51bbf9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.656782] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f589209-27bb-494a-9be1-1824c74505f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.672208] env[62066]: DEBUG nova.compute.manager [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Received event network-changed-eee10f4a-57b6-40ad-8049-87544bda2f33 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.673328] env[62066]: DEBUG nova.compute.manager [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Refreshing instance network info cache due to event network-changed-eee10f4a-57b6-40ad-8049-87544bda2f33. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.673328] env[62066]: DEBUG oslo_concurrency.lockutils [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] Acquiring lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.673328] env[62066]: DEBUG oslo_concurrency.lockutils [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] Acquired lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.673328] env[62066]: DEBUG nova.network.neutron [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Refreshing network info cache for port eee10f4a-57b6-40ad-8049-87544bda2f33 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 694.688484] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae818ddb-5fab-4356-b72d-61558d6f190a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.695484] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdd234b-6887-4e30-9c07-34a9e1e135c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.709436] env[62066]: DEBUG nova.virt.block_device [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating existing volume attachment record: b54fb348-249d-49e2-b05b-e94da5e9ef1e {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 694.806923] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.838712] env[62066]: ERROR nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 694.838712] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.838712] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.838712] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.838712] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.838712] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.838712] env[62066]: ERROR nova.compute.manager raise self.value [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.838712] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.838712] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.838712] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.839243] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.839243] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.839243] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 694.839243] env[62066]: ERROR nova.compute.manager [ 694.839860] env[62066]: Traceback (most recent call last): [ 694.839937] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.839937] env[62066]: listener.cb(fileno) [ 694.839937] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.839937] env[62066]: result = function(*args, **kwargs) [ 694.839937] env[62066]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.839937] env[62066]: return func(*args, **kwargs) [ 694.839937] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.839937] env[62066]: raise e [ 694.839937] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.839937] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 694.839937] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.839937] env[62066]: created_port_ids = self._update_ports_for_instance( [ 694.839937] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.839937] env[62066]: with excutils.save_and_reraise_exception(): [ 694.840364] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.840364] env[62066]: self.force_reraise() [ 694.840364] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.840364] env[62066]: raise self.value [ 694.840364] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.840364] env[62066]: updated_port = self._update_port( [ 694.840364] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.840364] env[62066]: _ensure_no_port_binding_failure(port) [ 694.840364] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.840364] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.840364] env[62066]: nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 694.840364] env[62066]: Removing descriptor: 18 [ 694.994817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.994817] env[62066]: ERROR nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Traceback (most recent call last): [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.driver.spawn(context, instance, image_meta, [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.994817] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] vm_ref = self.build_virtual_machine(instance, [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] for vif in network_info: [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self._sync_wrapper(fn, *args, **kwargs) [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.wait() [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self[:] = self._gt.wait() [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self._exit_event.wait() [ 694.995129] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] result = hub.switch() [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return self.greenlet.switch() [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] result = function(*args, **kwargs) [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] return func(*args, **kwargs) [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise e [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] nwinfo = self.network_api.allocate_for_instance( [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.995449] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] created_port_ids = self._update_ports_for_instance( [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] with excutils.save_and_reraise_exception(): [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] self.force_reraise() [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise self.value [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] updated_port = self._update_port( [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] _ensure_no_port_binding_failure(port) [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.995765] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] raise exception.PortBindingFailed(port_id=port['id']) [ 694.996071] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] nova.exception.PortBindingFailed: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. [ 694.996071] env[62066]: ERROR nova.compute.manager [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] [ 694.996071] env[62066]: DEBUG nova.compute.utils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 694.997908] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.216s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.000586] env[62066]: INFO nova.compute.claims [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.004643] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Build of instance e7b9c093-03a1-4cf9-bca5-41413aa4bcd8 was re-scheduled: Binding failed for port a29f546d-f1f4-4728-b265-084cd4781a8d, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.009014] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.009014] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquiring lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.009014] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Acquired lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.009014] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.021620] env[62066]: INFO nova.compute.manager [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Took 26.77 seconds to build instance. [ 695.199207] env[62066]: DEBUG nova.network.neutron [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.301553] env[62066]: DEBUG nova.network.neutron [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.342998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.528132] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d96ed8c7-438c-4d5c-9c1b-fbc1ec19b6b7 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.384s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.539352] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.615007] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.804439] env[62066]: DEBUG oslo_concurrency.lockutils [req-bb08d25e-312c-48af-b8bd-6e4097662582 req-1c036f9f-bbd7-4f22-9c0e-355eb796bd25 service nova] Releasing lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.030279] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.117349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Releasing lock "refresh_cache-e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.117608] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.117805] env[62066]: DEBUG nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.117977] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.138462] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.201672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "92b150af-351a-43ab-952a-209225733ac8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.201672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.228536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.228815] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.401463] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ed847c-b276-4a38-9f20-e395eeb78af3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.409765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8408dfd0-2ca6-43e5-ba28-acd834cdef43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.441103] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731c9b4e-b244-4704-a009-7cf1f89bae6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.448632] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27eb268-3097-45e8-9d90-9ceef123d6d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.463893] env[62066]: DEBUG nova.compute.provider_tree [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.478789] env[62066]: INFO nova.compute.manager [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Rebuilding instance [ 696.522731] env[62066]: DEBUG nova.compute.manager [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.523226] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220f41b9-b3c7-4264-990e-d0feeeaec445 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.562241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.643707] env[62066]: DEBUG nova.network.neutron [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.689362] env[62066]: DEBUG nova.compute.manager [req-e0f32cda-e2bf-497e-9491-3dda9eee5e1c req-4384668c-466c-4661-9c7c-d1e6e4641b56 service nova] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Received event network-vif-deleted-eee10f4a-57b6-40ad-8049-87544bda2f33 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.827121] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 696.827597] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.827808] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.827964] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.828167] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.828314] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.828459] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.828658] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.828859] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.829057] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.829227] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.829398] env[62066]: DEBUG nova.virt.hardware [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.830283] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe6c534-b1b2-42a4-a8f0-839f1ad10dcc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.838587] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc2f428-1cef-4b9a-bd24-5c75f0bf23b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.852339] env[62066]: ERROR nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Traceback (most recent call last): [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] yield resources [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.driver.spawn(context, instance, image_meta, [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] vm_ref = self.build_virtual_machine(instance, [ 696.852339] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] for vif in network_info: [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return self._sync_wrapper(fn, *args, **kwargs) [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.wait() [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self[:] = self._gt.wait() [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return self._exit_event.wait() [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.852673] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] current.throw(*self._exc) [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] result = function(*args, **kwargs) [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return func(*args, **kwargs) [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise e [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] nwinfo = self.network_api.allocate_for_instance( [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] created_port_ids = self._update_ports_for_instance( [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] with excutils.save_and_reraise_exception(): [ 696.853244] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.force_reraise() [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise self.value [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] updated_port = self._update_port( [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] _ensure_no_port_binding_failure(port) [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise exception.PortBindingFailed(port_id=port['id']) [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 696.853816] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] [ 696.853816] env[62066]: INFO nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Terminating instance [ 696.854779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquiring lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.854929] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquired lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.855106] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.967119] env[62066]: DEBUG nova.scheduler.client.report [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.034108] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 697.034423] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea749a88-7d12-462e-a681-a0a71400a748 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.042634] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 697.042634] env[62066]: value = "task-1155742" [ 697.042634] env[62066]: _type = "Task" [ 697.042634] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.053589] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.146467] env[62066]: INFO nova.compute.manager [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] [instance: e7b9c093-03a1-4cf9-bca5-41413aa4bcd8] Took 1.03 seconds to deallocate network for instance. [ 697.373518] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.454326] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.472515] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.473022] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.475426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.474s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.475646] env[62066]: DEBUG nova.objects.instance [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lazy-loading 'resources' on Instance uuid b4cb3d6d-d8ec-4632-8020-fae39b18143e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 697.552780] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155742, 'name': PowerOffVM_Task, 'duration_secs': 0.140044} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.552912] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 697.553102] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.553830] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d9cbf1-9dce-4c11-bc05-8cd4f706a22a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.560328] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 697.560542] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-656f6086-0a75-4f63-b20f-89f08cc7ee0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.588448] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 697.588664] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 697.588917] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Deleting the datastore file [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.589180] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a863cfa-a9ce-4e91-8a62-298f0d1f31f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.595265] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 697.595265] env[62066]: value = "task-1155744" [ 697.595265] env[62066]: _type = "Task" [ 697.595265] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.602583] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.959556] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Releasing lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.960163] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.960498] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96b86322-5e2a-4f13-b7bd-1160f5517ef7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.968815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ae14fb-f6fb-4d6c-adff-401b65458656 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.979901] env[62066]: DEBUG nova.compute.utils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 697.983833] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.984014] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 697.996478] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance ba33e43a-fbed-4a5a-9c05-a5a29082ebe9 could not be found. [ 697.996911] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.998807] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75bacfa7-2a35-411a-a3ab-076d77e5953a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.008715] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b87dfd-2386-4e8b-9c2e-8a125e345b31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.029314] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba33e43a-fbed-4a5a-9c05-a5a29082ebe9 could not be found. [ 698.029525] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.029713] env[62066]: INFO nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Took 0.07 seconds to destroy the instance on the hypervisor. [ 698.029998] env[62066]: DEBUG oslo.service.loopingcall [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.032323] env[62066]: DEBUG nova.compute.manager [-] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.032425] env[62066]: DEBUG nova.network.neutron [-] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.035961] env[62066]: DEBUG nova.policy [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dce5e9b642424ef5b202eb2295a3fe1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7dba0273591846a0a9522614b16bce62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.053451] env[62066]: DEBUG nova.network.neutron [-] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.104643] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100847} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.107294] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 698.107505] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 698.107706] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.174776] env[62066]: INFO nova.scheduler.client.report [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Deleted allocations for instance e7b9c093-03a1-4cf9-bca5-41413aa4bcd8 [ 698.327229] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Successfully created port: 4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.348733] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81700896-e314-49bd-882f-ceb57b9878b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.357839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647dcb70-abfe-440d-8c25-637d613e1eb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.388915] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeb9a33-fe76-486c-8639-62a9df6e9409 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.395864] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0a1a44-78b2-4a31-b64b-0096a551f767 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.409908] env[62066]: DEBUG nova.compute.provider_tree [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.486155] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.556409] env[62066]: DEBUG nova.network.neutron [-] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.682604] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a6b4d091-e4b4-4c10-b951-da7a7a0b0ee3 tempest-ServersAdminNegativeTestJSON-2147273100 tempest-ServersAdminNegativeTestJSON-2147273100-project-member] Lock "e7b9c093-03a1-4cf9-bca5-41413aa4bcd8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.956s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.913115] env[62066]: DEBUG nova.scheduler.client.report [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.060059] env[62066]: INFO nova.compute.manager [-] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Took 1.03 seconds to deallocate network for instance. [ 699.141297] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.141610] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.141801] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.141998] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.142158] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.142310] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.142517] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.142672] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.143160] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.143160] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.143160] env[62066]: DEBUG nova.virt.hardware [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.144034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd08f39-7d92-40f4-a208-05969e1da287 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.153277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40ea7f8-7a60-4049-8fda-43f656f34049 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.167526] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.172248] env[62066]: DEBUG oslo.service.loopingcall [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.173053] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.173053] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b7ddff6-bf33-4001-a2dd-4cdc8afdec6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.185567] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.189404] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.189404] env[62066]: value = "task-1155745" [ 699.189404] env[62066]: _type = "Task" [ 699.189404] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.198222] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155745, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.420454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.421818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.828s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.444946] env[62066]: INFO nova.scheduler.client.report [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Deleted allocations for instance b4cb3d6d-d8ec-4632-8020-fae39b18143e [ 699.495660] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.520585] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:53:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f58b5fca-fe02-4466-adf4-baeb8793402a',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-738942274',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.520693] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.520811] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.521018] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.521214] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.521393] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.521648] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.521845] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.522062] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.522238] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.522447] env[62066]: DEBUG nova.virt.hardware [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.523437] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e1f2be-af79-48ce-a347-09d30bb369d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.533110] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792db4ec-6da3-4313-b8c0-7515fa032540 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.617239] env[62066]: INFO nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Took 0.56 seconds to detach 1 volumes for instance. [ 699.619569] env[62066]: DEBUG nova.compute.claims [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.619790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.703253] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155745, 'name': CreateVM_Task, 'duration_secs': 0.299515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.703430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 699.703832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.704459] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.704459] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 699.704610] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e25dbf78-3830-4232-ba81-18e915566b27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.711345] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 699.711345] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5237d4e3-8759-0c72-e45a-fe7546ef3618" [ 699.711345] env[62066]: _type = "Task" [ 699.711345] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.713376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.718434] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5237d4e3-8759-0c72-e45a-fe7546ef3618, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.780210] env[62066]: DEBUG nova.compute.manager [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Received event network-vif-plugged-4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.780210] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] Acquiring lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.780210] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.780210] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.780210] env[62066]: DEBUG nova.compute.manager [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] No waiting events found dispatching network-vif-plugged-4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 699.780593] env[62066]: WARNING nova.compute.manager [req-97b6c4be-bd1a-4342-82a9-394a6cdf6f81 req-8194812d-0cd8-40c7-aadb-bfde62d00a17 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Received unexpected event network-vif-plugged-4627c549-cbf1-40f9-94bb-0f08255025ae for instance with vm_state building and task_state spawning. [ 699.885443] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Successfully updated port: 4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 699.954363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da091437-2041-4607-a184-a61b3497cd21 tempest-ServerShowV247Test-1295865255 tempest-ServerShowV247Test-1295865255-project-member] Lock "b4cb3d6d-d8ec-4632-8020-fae39b18143e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.670s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.223441] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5237d4e3-8759-0c72-e45a-fe7546ef3618, 'name': SearchDatastore_Task, 'duration_secs': 0.009538} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.223724] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.223935] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.224204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.224348] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.224517] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.224769] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa6d83f4-8326-471b-bb17-0eaadb408442 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.233550] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479475a0-2061-4212-b29d-825ace0bbfd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.238786] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.239101] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 700.240071] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-310c7caa-5a95-495d-ab39-bcc5f9d18be6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.245584] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db334f16-3125-4a43-ac85-1d4be8b2a5f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.250423] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 700.250423] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52528edc-6f57-c33d-bfd2-65ca8d499cdc" [ 700.250423] env[62066]: _type = "Task" [ 700.250423] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.278784] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e9e1c4-3185-4aa1-aeb0-2551d57239f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.285308] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52528edc-6f57-c33d-bfd2-65ca8d499cdc, 'name': SearchDatastore_Task, 'duration_secs': 0.008593} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.286504] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0450d47d-9767-4563-8646-b950cc27d9b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.292342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a49715-2f8f-442c-9c9d-49649754a4b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.298633] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 700.298633] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5227c8aa-6729-8a31-c878-3e3e4a398503" [ 700.298633] env[62066]: _type = "Task" [ 700.298633] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.308985] env[62066]: DEBUG nova.compute.provider_tree [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.314915] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5227c8aa-6729-8a31-c878-3e3e4a398503, 'name': SearchDatastore_Task, 'duration_secs': 0.008421} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.315166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.315416] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 700.315647] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77ebc98f-8b64-486d-8f96-17b60828b146 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.323098] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 700.323098] env[62066]: value = "task-1155746" [ 700.323098] env[62066]: _type = "Task" [ 700.323098] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.331236] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155746, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.387673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.387807] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.387965] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 700.815349] env[62066]: DEBUG nova.scheduler.client.report [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.834999] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155746, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482814} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.836036] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.836464] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.836815] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2031a2c3-ac26-407a-be23-bfb7828fb335 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.843626] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 700.843626] env[62066]: value = "task-1155747" [ 700.843626] env[62066]: _type = "Task" [ 700.843626] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.852653] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.953217] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.216396] env[62066]: DEBUG nova.network.neutron [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.318934] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.320994] env[62066]: ERROR nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Traceback (most recent call last): [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.driver.spawn(context, instance, image_meta, [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] vm_ref = self.build_virtual_machine(instance, [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.320994] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] for vif in network_info: [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return self._sync_wrapper(fn, *args, **kwargs) [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.wait() [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self[:] = self._gt.wait() [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return self._exit_event.wait() [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] current.throw(*self._exc) [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.321422] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] result = function(*args, **kwargs) [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] return func(*args, **kwargs) [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise e [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] nwinfo = self.network_api.allocate_for_instance( [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] created_port_ids = self._update_ports_for_instance( [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] with excutils.save_and_reraise_exception(): [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] self.force_reraise() [ 701.321798] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise self.value [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] updated_port = self._update_port( [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] _ensure_no_port_binding_failure(port) [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] raise exception.PortBindingFailed(port_id=port['id']) [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] nova.exception.PortBindingFailed: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. [ 701.322191] env[62066]: ERROR nova.compute.manager [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] [ 701.327052] env[62066]: DEBUG nova.compute.utils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 701.327052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.206s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.328458] env[62066]: INFO nova.compute.claims [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.334438] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Build of instance d584bc1e-b5a3-4885-a6fb-efec8267d894 was re-scheduled: Binding failed for port 291a9f0b-2e9f-488a-8cc2-c31edc469a7a, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.334883] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.335126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquiring lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.335273] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Acquired lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.335428] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.359541] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056881} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.360026] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.360766] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892479e3-ab3a-45b2-ae78-a16f825530ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.385668] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.385668] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73771002-8e7f-4625-b61f-cbce25b67721 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.405403] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 701.405403] env[62066]: value = "task-1155748" [ 701.405403] env[62066]: _type = "Task" [ 701.405403] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.413821] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.717646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.717775] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Instance network_info: |[{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 701.718225] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:3d:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4627c549-cbf1-40f9-94bb-0f08255025ae', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 701.725981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Creating folder: Project (7dba0273591846a0a9522614b16bce62). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.726273] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f71e4a99-53ea-47a0-8edc-4ef47b869f00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.736288] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Created folder: Project (7dba0273591846a0a9522614b16bce62) in parent group-v251573. [ 701.736480] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Creating folder: Instances. Parent ref: group-v251595. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.736800] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fbf771b6-11e6-44e8-a39e-f282dd64e59a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.745292] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Created folder: Instances in parent group-v251595. [ 701.745403] env[62066]: DEBUG oslo.service.loopingcall [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.745536] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 701.745731] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56647147-c549-4318-a706-32e34f7d2c62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.765175] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 701.765175] env[62066]: value = "task-1155751" [ 701.765175] env[62066]: _type = "Task" [ 701.765175] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.774843] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155751, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.863374] env[62066]: DEBUG nova.compute.manager [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Received event network-changed-4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.863588] env[62066]: DEBUG nova.compute.manager [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Refreshing instance network info cache due to event network-changed-4627c549-cbf1-40f9-94bb-0f08255025ae. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.863853] env[62066]: DEBUG oslo_concurrency.lockutils [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.864610] env[62066]: DEBUG oslo_concurrency.lockutils [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.865268] env[62066]: DEBUG nova.network.neutron [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Refreshing network info cache for port 4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.870271] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.921312] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155748, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.994444] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.228737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "d9163e64-309b-4381-8819-15757f83ac2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.229702] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.275879] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155751, 'name': CreateVM_Task, 'duration_secs': 0.336274} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.275879] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 702.284154] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.284154] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.284542] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 702.284799] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e129b9ee-2e38-4100-b0f2-657add05cd4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.290154] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 702.290154] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528d3bb0-fb66-605b-52dc-ab3ba444cfd3" [ 702.290154] env[62066]: _type = "Task" [ 702.290154] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.297907] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528d3bb0-fb66-605b-52dc-ab3ba444cfd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.419500] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155748, 'name': ReconfigVM_Task, 'duration_secs': 0.769227} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.419781] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Reconfigured VM instance instance-00000026 to attach disk [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb/d6837004-9b24-45f7-9ed7-48c4503859cb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.420395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8470e84f-260a-4bef-a76e-27bdc20619c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.427371] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 702.427371] env[62066]: value = "task-1155752" [ 702.427371] env[62066]: _type = "Task" [ 702.427371] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.436168] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155752, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.499026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Releasing lock "refresh_cache-d584bc1e-b5a3-4885-a6fb-efec8267d894" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.499026] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.499026] env[62066]: DEBUG nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.499026] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.518211] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.654330] env[62066]: DEBUG nova.network.neutron [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updated VIF entry in instance network info cache for port 4627c549-cbf1-40f9-94bb-0f08255025ae. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 702.654682] env[62066]: DEBUG nova.network.neutron [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.771286] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6fa093-cb63-4fc1-84d0-d31925e24176 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.779314] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70de6b28-0ead-4dd9-a145-4dd321863a8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.815284] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14dc462-1a1e-4c0d-9075-1baca49bcd0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.823018] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528d3bb0-fb66-605b-52dc-ab3ba444cfd3, 'name': SearchDatastore_Task, 'duration_secs': 0.008416} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.825106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.825382] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 702.825661] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.825841] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.826066] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 702.826366] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b96077d-ca1e-4ee4-abf4-df061cef36c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.829012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bb8da6-14bf-4ec9-9de5-8f5c54762e9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.842497] env[62066]: DEBUG nova.compute.provider_tree [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.847133] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 702.847133] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 702.847133] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4598ee16-45c2-4115-9a0b-ac99dd32a71f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.851471] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 702.851471] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5268de3e-9396-3fbd-f861-5c0708decef7" [ 702.851471] env[62066]: _type = "Task" [ 702.851471] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.860221] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268de3e-9396-3fbd-f861-5c0708decef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.936938] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155752, 'name': Rename_Task, 'duration_secs': 0.128272} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.937799] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.937799] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2af70a1-aaad-4e9d-9c2a-2dc2aaaa50bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.943944] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 702.943944] env[62066]: value = "task-1155753" [ 702.943944] env[62066]: _type = "Task" [ 702.943944] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.951106] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.020609] env[62066]: DEBUG nova.network.neutron [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.160915] env[62066]: DEBUG oslo_concurrency.lockutils [req-3724eb1c-c7ce-4a9b-9869-f3375208cd1a req-b88c1c39-1513-4c82-9587-42a3451eef69 service nova] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.240843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.241031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.345105] env[62066]: DEBUG nova.scheduler.client.report [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.364676] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268de3e-9396-3fbd-f861-5c0708decef7, 'name': SearchDatastore_Task, 'duration_secs': 0.007639} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.365564] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-857d3990-edc7-4b37-8926-888d5afbd568 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.372327] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 703.372327] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52464e00-2906-0d83-937e-22c049663d4e" [ 703.372327] env[62066]: _type = "Task" [ 703.372327] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.380585] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52464e00-2906-0d83-937e-22c049663d4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.453233] env[62066]: DEBUG oslo_vmware.api [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155753, 'name': PowerOnVM_Task, 'duration_secs': 0.450387} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.453520] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.453722] env[62066]: DEBUG nova.compute.manager [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 703.454481] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb8431c-397d-4d11-b53f-aa8c298493e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.525009] env[62066]: INFO nova.compute.manager [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] [instance: d584bc1e-b5a3-4885-a6fb-efec8267d894] Took 1.03 seconds to deallocate network for instance. [ 703.849893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.850439] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.853034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.796s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.858763] env[62066]: INFO nova.compute.claims [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.889558] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52464e00-2906-0d83-937e-22c049663d4e, 'name': SearchDatastore_Task, 'duration_secs': 0.013113} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.890033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.890550] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 703.891494] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9125c15-e37b-4a23-b0fa-681ae1c880ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.900904] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 703.900904] env[62066]: value = "task-1155754" [ 703.900904] env[62066]: _type = "Task" [ 703.900904] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.909376] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155754, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.971885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.323380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "d6837004-9b24-45f7-9ed7-48c4503859cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.323727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.323998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "d6837004-9b24-45f7-9ed7-48c4503859cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.324470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.324470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.327292] env[62066]: INFO nova.compute.manager [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Terminating instance [ 704.332082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.332177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquired lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.332460] env[62066]: DEBUG nova.network.neutron [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.366379] env[62066]: DEBUG nova.compute.utils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.369853] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.371372] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.410692] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155754, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465811} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.411466] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 704.411466] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 704.411466] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96c106a3-6203-46aa-bf44-29db66fc4bb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.415027] env[62066]: DEBUG nova.policy [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.418263] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 704.418263] env[62066]: value = "task-1155755" [ 704.418263] env[62066]: _type = "Task" [ 704.418263] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.426115] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155755, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.558054] env[62066]: INFO nova.scheduler.client.report [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Deleted allocations for instance d584bc1e-b5a3-4885-a6fb-efec8267d894 [ 704.686870] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Successfully created port: 59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.855952] env[62066]: DEBUG nova.network.neutron [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.873118] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.932845] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155755, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.933121] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 704.933895] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3edd31-c44d-4e7a-8e3d-7c7a7f57a4cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.939578] env[62066]: DEBUG nova.network.neutron [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.967410] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 704.972442] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3bc0219-77e0-4829-9e84-666d64ceeb05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.993692] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 704.993692] env[62066]: value = "task-1155756" [ 704.993692] env[62066]: _type = "Task" [ 704.993692] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.006095] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155756, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.066343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-233062b9-585f-4588-bba0-bc171f8bc567 tempest-AttachInterfacesV270Test-777259799 tempest-AttachInterfacesV270Test-777259799-project-member] Lock "d584bc1e-b5a3-4885-a6fb-efec8267d894" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.281s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.245820] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a7c5f5-0607-4973-a10c-ad31354ac078 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.253739] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad7b141-bd79-4d3e-b53d-8b3a013a6eaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.284711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce42da7-7b29-4380-bd85-85b0bc2fc097 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.292573] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bedcaa-dfc6-4997-9923-83f2bd567d99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.306363] env[62066]: DEBUG nova.compute.provider_tree [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.443577] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Releasing lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.443680] env[62066]: DEBUG nova.compute.manager [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.443875] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.444754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed296f63-6492-463f-9fa7-739077d48e83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.452613] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 705.452855] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69654d35-326b-46e1-939a-cd06fb9e46ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.459229] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 705.459229] env[62066]: value = "task-1155757" [ 705.459229] env[62066]: _type = "Task" [ 705.459229] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.467019] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.504037] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155756, 'name': ReconfigVM_Task, 'duration_secs': 0.32275} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.504091] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 705.504710] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c157230-91b3-40f6-8bed-2dc0bf133d8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.510631] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 705.510631] env[62066]: value = "task-1155758" [ 705.510631] env[62066]: _type = "Task" [ 705.510631] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.519165] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155758, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.568596] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.809564] env[62066]: DEBUG nova.scheduler.client.report [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.888025] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.912019] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.912019] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.912019] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.912019] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.912278] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.912278] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.912278] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.912278] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.912475] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.912635] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.912800] env[62066]: DEBUG nova.virt.hardware [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.913686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d3bf29-c74f-492d-a6a8-d1cb10b8de02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.921649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6eca85-b700-4f6e-a028-b91694931d72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.967883] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155757, 'name': PowerOffVM_Task, 'duration_secs': 0.162694} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.968162] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 705.968337] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 705.968593] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4b549df-3e24-46f6-b955-d141b239865c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.999715] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 705.999923] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.000098] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Deleting the datastore file [datastore2] d6837004-9b24-45f7-9ed7-48c4503859cb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.000352] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-969bab25-63ae-453e-9997-7c4032763485 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.006841] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for the task: (returnval){ [ 706.006841] env[62066]: value = "task-1155760" [ 706.006841] env[62066]: _type = "Task" [ 706.006841] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.017921] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.022828] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155758, 'name': Rename_Task, 'duration_secs': 0.152359} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.023123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 706.023364] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22e71983-465f-4c1d-a8c4-60d5c4a67d5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.028775] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 706.028775] env[62066]: value = "task-1155761" [ 706.028775] env[62066]: _type = "Task" [ 706.028775] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.036466] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.095438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.179297] env[62066]: DEBUG nova.compute.manager [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Received event network-vif-plugged-59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.179556] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] Acquiring lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.179780] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.179974] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.180495] env[62066]: DEBUG nova.compute.manager [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] No waiting events found dispatching network-vif-plugged-59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 706.180696] env[62066]: WARNING nova.compute.manager [req-f4f1be98-64b6-4cee-96b8-d100f5b4dea7 req-9f1c18ad-21a4-4c63-8d99-d76ee9469c33 service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Received unexpected event network-vif-plugged-59b6e38f-b186-420c-89c8-c860e8310108 for instance with vm_state building and task_state spawning. [ 706.242871] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Successfully updated port: 59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.316213] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.316858] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.319379] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.196s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.322759] env[62066]: INFO nova.compute.claims [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.520279] env[62066]: DEBUG oslo_vmware.api [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Task: {'id': task-1155760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.520641] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 706.520899] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 706.521306] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.521596] env[62066]: INFO nova.compute.manager [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Took 1.08 seconds to destroy the instance on the hypervisor. [ 706.521958] env[62066]: DEBUG oslo.service.loopingcall [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.522241] env[62066]: DEBUG nova.compute.manager [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.522428] env[62066]: DEBUG nova.network.neutron [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.539198] env[62066]: DEBUG oslo_vmware.api [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155761, 'name': PowerOnVM_Task, 'duration_secs': 0.480853} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.539515] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 706.539749] env[62066]: INFO nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Took 7.04 seconds to spawn the instance on the hypervisor. [ 706.540135] env[62066]: DEBUG nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 706.542287] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcc92c5-946f-4dbb-8fa4-f18f352c6512 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.545293] env[62066]: DEBUG nova.network.neutron [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.752987] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.753260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.753475] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.827860] env[62066]: DEBUG nova.compute.utils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 706.830192] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.830427] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 707.023532] env[62066]: DEBUG nova.policy [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '232c6acdc3d444bca17927d6fb209670', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90669b8114cb42cdbac064f40b735c90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.049399] env[62066]: DEBUG nova.network.neutron [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.066127] env[62066]: INFO nova.compute.manager [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Took 29.30 seconds to build instance. [ 707.318302] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.337696] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.554632] env[62066]: INFO nova.compute.manager [-] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Took 1.03 seconds to deallocate network for instance. [ 707.560729] env[62066]: DEBUG nova.network.neutron [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updating instance_info_cache with network_info: [{"id": "59b6e38f-b186-420c-89c8-c860e8310108", "address": "fa:16:3e:5c:97:e8", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b6e38f-b1", "ovs_interfaceid": "59b6e38f-b186-420c-89c8-c860e8310108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.569513] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5c09010-f0d6-4051-b2fc-c1e6ecaef2d1 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.546s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.616949] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Successfully created port: ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.720815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92eb3e1-5e3b-4c5d-8118-63404c2f5b02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.733483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d71ece-b084-4538-b170-6270f5d403c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.768019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbd0361-cecd-4c90-a142-50e3a79a1e58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.776119] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5088f5-6d22-4410-b4b1-be1d294d4cd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.801065] env[62066]: DEBUG nova.compute.provider_tree [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.064336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.064753] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.065798] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Instance network_info: |[{"id": "59b6e38f-b186-420c-89c8-c860e8310108", "address": "fa:16:3e:5c:97:e8", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b6e38f-b1", "ovs_interfaceid": "59b6e38f-b186-420c-89c8-c860e8310108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.066978] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:97:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59b6e38f-b186-420c-89c8-c860e8310108', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.075089] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating folder: Project (1160432c71b042efa6c0e45cf58b37cb). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.075478] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.077869] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ee302e6-043f-4d6c-abc2-5f1e905b51c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.092742] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created folder: Project (1160432c71b042efa6c0e45cf58b37cb) in parent group-v251573. [ 708.092938] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating folder: Instances. Parent ref: group-v251598. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.093249] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f085d40b-d465-4815-8d94-c78a66864df6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.101929] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created folder: Instances in parent group-v251598. [ 708.102187] env[62066]: DEBUG oslo.service.loopingcall [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.102381] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.102586] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a985fb6-8bc5-4fd0-a16f-2600966f9c5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.122668] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.122668] env[62066]: value = "task-1155764" [ 708.122668] env[62066]: _type = "Task" [ 708.122668] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.130579] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155764, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.281728] env[62066]: DEBUG nova.compute.manager [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Received event network-changed-59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.281728] env[62066]: DEBUG nova.compute.manager [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Refreshing instance network info cache due to event network-changed-59b6e38f-b186-420c-89c8-c860e8310108. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.281873] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] Acquiring lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.282147] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] Acquired lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.282387] env[62066]: DEBUG nova.network.neutron [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Refreshing network info cache for port 59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.305600] env[62066]: DEBUG nova.scheduler.client.report [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.346429] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.377191] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.377535] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.377629] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.377803] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.377962] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.378192] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.378463] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.378641] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.378846] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.379139] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.379385] env[62066]: DEBUG nova.virt.hardware [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.380554] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34447bd6-695c-4be5-a2f5-087000f4d1fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.389239] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85472657-32d8-4ef6-8336-95e7463c4961 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.594516] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.636143] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155764, 'name': CreateVM_Task, 'duration_secs': 0.320055} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.636430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 708.637407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.637681] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.638251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 708.638604] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb38670e-7d8f-45b5-aa96-ca3c00a792ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.643656] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 708.643656] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cd16-e94f-19b7-d011-1f9614b8ba7c" [ 708.643656] env[62066]: _type = "Task" [ 708.643656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.651394] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cd16-e94f-19b7-d011-1f9614b8ba7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.812675] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.814027] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.815797] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.793s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.817324] env[62066]: INFO nova.compute.claims [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.076543] env[62066]: DEBUG nova.network.neutron [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updated VIF entry in instance network info cache for port 59b6e38f-b186-420c-89c8-c860e8310108. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 709.077508] env[62066]: DEBUG nova.network.neutron [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updating instance_info_cache with network_info: [{"id": "59b6e38f-b186-420c-89c8-c860e8310108", "address": "fa:16:3e:5c:97:e8", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b6e38f-b1", "ovs_interfaceid": "59b6e38f-b186-420c-89c8-c860e8310108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.155527] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cd16-e94f-19b7-d011-1f9614b8ba7c, 'name': SearchDatastore_Task, 'duration_secs': 0.00933} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.156023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.156550] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.157073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.157375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.157702] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.158802] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fec76f6-4b64-4680-9e48-33fec82ffc38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.167035] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.167472] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 709.168065] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84ea788b-cbfa-4a98-bc44-baeda8a4d3e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.173441] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 709.173441] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523edb76-8cd8-9a43-f744-4935beecf291" [ 709.173441] env[62066]: _type = "Task" [ 709.173441] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.181983] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523edb76-8cd8-9a43-f744-4935beecf291, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.234505] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Successfully updated port: ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 709.288315] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.288546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.321384] env[62066]: DEBUG nova.compute.utils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 709.324543] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.324543] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 709.378446] env[62066]: DEBUG nova.policy [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '232c6acdc3d444bca17927d6fb209670', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90669b8114cb42cdbac064f40b735c90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.581295] env[62066]: DEBUG oslo_concurrency.lockutils [req-f4095226-c1dd-499a-8837-2cc67c196ef5 req-c0130140-8b13-4d31-8999-9a85c5e6f2af service nova] Releasing lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.624470] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Successfully created port: 1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.685569] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523edb76-8cd8-9a43-f744-4935beecf291, 'name': SearchDatastore_Task, 'duration_secs': 0.008329} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.686361] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18614b5d-2f7d-4356-872b-395d8d8710ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.695597] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 709.695597] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5245714b-b77c-4810-786d-6da96a2d4059" [ 709.695597] env[62066]: _type = "Task" [ 709.695597] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.696237] env[62066]: DEBUG nova.compute.manager [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 709.717185] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5245714b-b77c-4810-786d-6da96a2d4059, 'name': SearchDatastore_Task, 'duration_secs': 0.008405} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.717658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.718078] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 212dac6a-a291-4ca8-87fb-97ebcca7976c/212dac6a-a291-4ca8-87fb-97ebcca7976c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.718669] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7449e306-7109-48ae-a4b2-c2529d812fd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.729627] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 709.729627] env[62066]: value = "task-1155765" [ 709.729627] env[62066]: _type = "Task" [ 709.729627] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.738499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.738499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.738663] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.741113] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.827902] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.217794] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.229140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3e77c1-673f-4e3c-8d9d-5f6b138bf9e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.243287] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465814} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.243914] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3724d8b8-5d88-4469-94eb-bc2eb7c5ff91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.247107] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 212dac6a-a291-4ca8-87fb-97ebcca7976c/212dac6a-a291-4ca8-87fb-97ebcca7976c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 710.247341] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.247808] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd99e747-bf1c-44c7-9ef6-7d498e4c218d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.279584] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdf22c6-7c3d-4581-a6dd-af1d199c6337 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.281981] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 710.281981] env[62066]: value = "task-1155766" [ 710.281981] env[62066]: _type = "Task" [ 710.281981] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.282820] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.292052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fc61b5-c5d5-47e9-8008-1db70007c63c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.298868] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.302176] env[62066]: DEBUG nova.compute.manager [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Received event network-vif-plugged-ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.302373] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Acquiring lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.302572] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.302734] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.302907] env[62066]: DEBUG nova.compute.manager [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] No waiting events found dispatching network-vif-plugged-ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 710.303128] env[62066]: WARNING nova.compute.manager [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Received unexpected event network-vif-plugged-ef7a5281-027a-42e2-b91c-21fc55677027 for instance with vm_state building and task_state spawning. [ 710.303289] env[62066]: DEBUG nova.compute.manager [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Received event network-changed-ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.303469] env[62066]: DEBUG nova.compute.manager [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Refreshing instance network info cache due to event network-changed-ef7a5281-027a-42e2-b91c-21fc55677027. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.303709] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Acquiring lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.312368] env[62066]: DEBUG nova.compute.provider_tree [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.436015] env[62066]: DEBUG nova.network.neutron [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Updating instance_info_cache with network_info: [{"id": "ef7a5281-027a-42e2-b91c-21fc55677027", "address": "fa:16:3e:30:7c:8a", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef7a5281-02", "ovs_interfaceid": "ef7a5281-027a-42e2-b91c-21fc55677027", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.794965] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.344098} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.795276] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.796217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b38620-ad0c-4268-8ea4-82fe89d7f97d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.818268] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 212dac6a-a291-4ca8-87fb-97ebcca7976c/212dac6a-a291-4ca8-87fb-97ebcca7976c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.819163] env[62066]: DEBUG nova.scheduler.client.report [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.823038] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b4bf5ce-5874-4b61-bb63-5c4fe3eb17fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.837292] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.021s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.837767] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.840614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.498s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.842249] env[62066]: INFO nova.compute.claims [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.846216] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.850924] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 710.850924] env[62066]: value = "task-1155767" [ 710.850924] env[62066]: _type = "Task" [ 710.850924] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.859834] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155767, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.872278] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.872535] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.872689] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.872864] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.873023] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.873169] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.873370] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.873525] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.873762] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.874025] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.874325] env[62066]: DEBUG nova.virt.hardware [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.875346] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da032d8c-1f9f-4d6b-a401-3aeb2e69b491 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.883926] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70992f7-aa4a-4c20-9a5a-6821f2032b71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.939039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.939529] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Instance network_info: |[{"id": "ef7a5281-027a-42e2-b91c-21fc55677027", "address": "fa:16:3e:30:7c:8a", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef7a5281-02", "ovs_interfaceid": "ef7a5281-027a-42e2-b91c-21fc55677027", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 710.939971] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Acquired lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.940302] env[62066]: DEBUG nova.network.neutron [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Refreshing network info cache for port ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.942028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:7c:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef7a5281-027a-42e2-b91c-21fc55677027', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.949590] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating folder: Project (90669b8114cb42cdbac064f40b735c90). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.952766] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5fcb80f-e0ee-46fc-8e43-1a82a6b37532 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.965014] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Created folder: Project (90669b8114cb42cdbac064f40b735c90) in parent group-v251573. [ 710.965227] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating folder: Instances. Parent ref: group-v251601. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.965471] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2011966f-a546-4d9a-87b3-f040db18fcca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.974260] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Created folder: Instances in parent group-v251601. [ 710.974486] env[62066]: DEBUG oslo.service.loopingcall [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.974775] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 710.974872] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07246ab5-78e6-42b9-8d8a-613daf7ae24f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.996547] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.996547] env[62066]: value = "task-1155770" [ 710.996547] env[62066]: _type = "Task" [ 710.996547] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.006187] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155770, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.277615] env[62066]: DEBUG nova.network.neutron [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Updated VIF entry in instance network info cache for port ef7a5281-027a-42e2-b91c-21fc55677027. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 711.277973] env[62066]: DEBUG nova.network.neutron [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Updating instance_info_cache with network_info: [{"id": "ef7a5281-027a-42e2-b91c-21fc55677027", "address": "fa:16:3e:30:7c:8a", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef7a5281-02", "ovs_interfaceid": "ef7a5281-027a-42e2-b91c-21fc55677027", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.290754] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Successfully updated port: 1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.346779] env[62066]: DEBUG nova.compute.utils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 711.350124] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.350299] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.363172] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155767, 'name': ReconfigVM_Task, 'duration_secs': 0.507326} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.363172] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 212dac6a-a291-4ca8-87fb-97ebcca7976c/212dac6a-a291-4ca8-87fb-97ebcca7976c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 711.363172] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2f0c939-2e4d-450c-b9fa-898b76180e8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.370036] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 711.370036] env[62066]: value = "task-1155771" [ 711.370036] env[62066]: _type = "Task" [ 711.370036] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.379731] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155771, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.396495] env[62066]: DEBUG nova.policy [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.507210] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155770, 'name': CreateVM_Task, 'duration_secs': 0.356066} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.507405] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 711.508089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.508482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.508552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.508865] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56c08f41-334f-44c9-9545-c1240a34cd57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.513365] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 711.513365] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d74610-451c-c0be-0908-3b23f28378a1" [ 711.513365] env[62066]: _type = "Task" [ 711.513365] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.520866] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d74610-451c-c0be-0908-3b23f28378a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.687008] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Successfully created port: 1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.780463] env[62066]: DEBUG oslo_concurrency.lockutils [req-858d562e-fdec-4fbf-a39d-96f2f22a8509 req-a57070b8-456c-4734-a2a9-90faaa0995fb service nova] Releasing lock "refresh_cache-01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.794839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.794839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.794839] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.851255] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.883600] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155771, 'name': Rename_Task, 'duration_secs': 0.296713} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.883903] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 711.884158] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a40d8a8c-5398-4a5a-bee4-ffe9d6e30053 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.894904] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 711.894904] env[62066]: value = "task-1155772" [ 711.894904] env[62066]: _type = "Task" [ 711.894904] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.903098] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.025853] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d74610-451c-c0be-0908-3b23f28378a1, 'name': SearchDatastore_Task, 'duration_secs': 0.018089} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.028277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.028509] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.028733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.028872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.029085] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.029722] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9e0d0e0-c5ff-407a-841e-a38bf33dead6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.037816] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.037816] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 712.038702] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a17f3da-5849-4914-bbc3-ab4295954676 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.045795] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 712.045795] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5255c4a0-fddf-66e0-cb3c-412ce8438696" [ 712.045795] env[62066]: _type = "Task" [ 712.045795] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.053540] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5255c4a0-fddf-66e0-cb3c-412ce8438696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.210375] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6828596-41d0-4f7d-bf76-923c0ff76a62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.217930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd1e84c-b681-4f97-8f11-7a1c8a445ad8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.250170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af298663-cb7e-4a3c-8531-640e27fb4589 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.258978] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08562b0a-0773-4ee4-943f-8921b3ac742a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.272998] env[62066]: DEBUG nova.compute.provider_tree [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.332675] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.363925] env[62066]: DEBUG nova.compute.manager [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Received event network-vif-plugged-1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.364202] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Acquiring lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.366017] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.366017] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.366017] env[62066]: DEBUG nova.compute.manager [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] No waiting events found dispatching network-vif-plugged-1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 712.366017] env[62066]: WARNING nova.compute.manager [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Received unexpected event network-vif-plugged-1cfa47b1-d2b5-4417-9791-d751e1c777fe for instance with vm_state building and task_state spawning. [ 712.366017] env[62066]: DEBUG nova.compute.manager [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Received event network-changed-1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.366227] env[62066]: DEBUG nova.compute.manager [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Refreshing instance network info cache due to event network-changed-1cfa47b1-d2b5-4417-9791-d751e1c777fe. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.366227] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Acquiring lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.406490] env[62066]: DEBUG oslo_vmware.api [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1155772, 'name': PowerOnVM_Task, 'duration_secs': 0.431185} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.406742] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 712.406928] env[62066]: INFO nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Took 6.52 seconds to spawn the instance on the hypervisor. [ 712.407112] env[62066]: DEBUG nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 712.407881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835e9ec7-d865-4149-a778-47e4803ef8d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.472391] env[62066]: DEBUG nova.network.neutron [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Updating instance_info_cache with network_info: [{"id": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "address": "fa:16:3e:20:41:a3", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cfa47b1-d2", "ovs_interfaceid": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.556263] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5255c4a0-fddf-66e0-cb3c-412ce8438696, 'name': SearchDatastore_Task, 'duration_secs': 0.008605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.556814] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9b28c79-5b3e-43d8-84ec-998462f68a8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.562200] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 712.562200] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528a02d5-614d-7434-8d50-cb98af370fec" [ 712.562200] env[62066]: _type = "Task" [ 712.562200] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.570693] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528a02d5-614d-7434-8d50-cb98af370fec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.776714] env[62066]: DEBUG nova.scheduler.client.report [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.863279] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.884864] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.885132] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.885293] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.885474] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.885615] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.885755] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.885956] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.886119] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.886280] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.886990] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.886990] env[62066]: DEBUG nova.virt.hardware [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.887457] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920589a5-3788-437e-83d4-e2894b308479 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.895241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd7f27b-9606-478d-bc80-a42af01f383a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.922898] env[62066]: INFO nova.compute.manager [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Took 30.82 seconds to build instance. [ 712.974722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.975009] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Instance network_info: |[{"id": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "address": "fa:16:3e:20:41:a3", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cfa47b1-d2", "ovs_interfaceid": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 712.975321] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Acquired lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.975643] env[62066]: DEBUG nova.network.neutron [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Refreshing network info cache for port 1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 712.977914] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:41:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d31a554-a94c-4471-892f-f65aa87b8279', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cfa47b1-d2b5-4417-9791-d751e1c777fe', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.984442] env[62066]: DEBUG oslo.service.loopingcall [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.985062] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 712.985261] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a29bdf9-4de8-4627-916a-0cfe6de37c7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.007687] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.007687] env[62066]: value = "task-1155773" [ 713.007687] env[62066]: _type = "Task" [ 713.007687] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.015835] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155773, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.075245] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528a02d5-614d-7434-8d50-cb98af370fec, 'name': SearchDatastore_Task, 'duration_secs': 0.011657} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.075846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.076166] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf/01e1df17-4b9d-4e12-bf6b-50b39c08bfbf.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 713.076432] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6031278b-ccb9-4e0b-a133-6fcd0a8c6790 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.082585] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 713.082585] env[62066]: value = "task-1155774" [ 713.082585] env[62066]: _type = "Task" [ 713.082585] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.091437] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.193314] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Successfully updated port: 1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 713.282386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.282746] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.285409] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.724s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.287385] env[62066]: INFO nova.compute.claims [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.425113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db230f80-9384-4940-a6a8-e3984de71880 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.741s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.525168] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155773, 'name': CreateVM_Task, 'duration_secs': 0.301502} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.527596] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 713.528327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.528473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.528790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 713.529412] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d95236d-3186-4dbe-8556-fe935c109888 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.536640] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 713.536640] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52658cef-7797-c71d-d90f-c32d706a48ad" [ 713.536640] env[62066]: _type = "Task" [ 713.536640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.547035] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52658cef-7797-c71d-d90f-c32d706a48ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.593876] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490969} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.594147] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf/01e1df17-4b9d-4e12-bf6b-50b39c08bfbf.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 713.594365] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.594608] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d400076-9ed5-4589-bc5e-e3c1ba60c3ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.600902] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 713.600902] env[62066]: value = "task-1155775" [ 713.600902] env[62066]: _type = "Task" [ 713.600902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.608665] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.695117] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.695287] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.695463] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.706980] env[62066]: DEBUG nova.network.neutron [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Updated VIF entry in instance network info cache for port 1cfa47b1-d2b5-4417-9791-d751e1c777fe. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 713.707368] env[62066]: DEBUG nova.network.neutron [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Updating instance_info_cache with network_info: [{"id": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "address": "fa:16:3e:20:41:a3", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cfa47b1-d2", "ovs_interfaceid": "1cfa47b1-d2b5-4417-9791-d751e1c777fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.794051] env[62066]: DEBUG nova.compute.utils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 713.798977] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.798977] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.835986] env[62066]: DEBUG nova.policy [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315ec256cd6b422a90f2914175cb49bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c93c0e92cfec42f4b4a20e9fb4a32088', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.928475] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.934380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "43d64f76-7f7a-4b95-b9df-c95218612998" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.934543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.051559] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52658cef-7797-c71d-d90f-c32d706a48ad, 'name': SearchDatastore_Task, 'duration_secs': 0.027135} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.051869] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.052107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.052339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.052474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.052644] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.053780] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09b55e5b-d8a4-4e49-98fa-699bced71a30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.063020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 714.063020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 714.063020] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4763934-3176-4ce1-ac2b-0784940bf635 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.068652] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 714.068652] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5216d664-d547-8cdb-a199-635f04650d99" [ 714.068652] env[62066]: _type = "Task" [ 714.068652] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.077355] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5216d664-d547-8cdb-a199-635f04650d99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.097666] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Successfully created port: 3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.110576] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260012} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.110890] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.112365] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37843bb8-f565-4f4f-b962-571fca0debbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.134998] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf/01e1df17-4b9d-4e12-bf6b-50b39c08bfbf.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.135353] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6f6a96b-3b12-468d-812f-789bbe5202a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.156548] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 714.156548] env[62066]: value = "task-1155776" [ 714.156548] env[62066]: _type = "Task" [ 714.156548] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.166866] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.209633] env[62066]: DEBUG oslo_concurrency.lockutils [req-762fc49c-1fcc-472e-873a-43b26de67f84 req-a0153099-f11b-41b0-b896-3367b19aa5b9 service nova] Releasing lock "refresh_cache-2f186d0a-91a0-4dc2-83bd-511099445af7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.243178] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.298302] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.400666] env[62066]: DEBUG nova.compute.manager [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-vif-plugged-1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.400895] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.401129] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.401342] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.401554] env[62066]: DEBUG nova.compute.manager [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] No waiting events found dispatching network-vif-plugged-1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 714.401686] env[62066]: WARNING nova.compute.manager [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received unexpected event network-vif-plugged-1436b8c7-b051-48ad-962f-30ccdc95b36b for instance with vm_state building and task_state spawning. [ 714.401839] env[62066]: DEBUG nova.compute.manager [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-changed-1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.401989] env[62066]: DEBUG nova.compute.manager [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing instance network info cache due to event network-changed-1436b8c7-b051-48ad-962f-30ccdc95b36b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 714.402207] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.453990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.472225] env[62066]: DEBUG nova.network.neutron [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.580870] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5216d664-d547-8cdb-a199-635f04650d99, 'name': SearchDatastore_Task, 'duration_secs': 0.013637} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.582992] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-649f8b17-e870-4f74-bf13-624d03a86c37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.587773] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 714.587773] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524e6637-331e-3065-9e08-99de5ec4b574" [ 714.587773] env[62066]: _type = "Task" [ 714.587773] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.597165] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524e6637-331e-3065-9e08-99de5ec4b574, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.664909] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155776, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.666361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9347aab-a554-44c3-8db5-e3eca4e386a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.672747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41149f6-dc55-4246-bb87-ecfd81f34bb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.701667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5799698-d536-49cb-ae57-56f3ceab30be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.708562] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9befb92-bb54-4547-84bb-d7cc83e75b75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.722304] env[62066]: DEBUG nova.compute.provider_tree [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.977193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.977193] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Instance network_info: |[{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 714.977615] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.977615] env[62066]: DEBUG nova.network.neutron [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing network info cache for port 1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.978427] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:01:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '65497291-07f3-434c-bd42-657a0cb03365', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1436b8c7-b051-48ad-962f-30ccdc95b36b', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.986551] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating folder: Project (42219a58a1514265b9d0b515eb517933). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.987686] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf1a52fd-7700-485a-badf-4b900e6f0011 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.999388] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created folder: Project (42219a58a1514265b9d0b515eb517933) in parent group-v251573. [ 714.999616] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating folder: Instances. Parent ref: group-v251605. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.999873] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef5ef02a-0ea2-48d8-834a-e22f27d2b40d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.009582] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created folder: Instances in parent group-v251605. [ 715.009823] env[62066]: DEBUG oslo.service.loopingcall [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.010019] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 715.010453] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3f1e424-58dd-480e-81d2-4a9757f223ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.029370] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.029370] env[62066]: value = "task-1155779" [ 715.029370] env[62066]: _type = "Task" [ 715.029370] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.038861] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155779, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.101594] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524e6637-331e-3065-9e08-99de5ec4b574, 'name': SearchDatastore_Task, 'duration_secs': 0.046701} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.101993] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.102629] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2f186d0a-91a0-4dc2-83bd-511099445af7/2f186d0a-91a0-4dc2-83bd-511099445af7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 715.102629] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55de9b61-1c42-45ae-9693-c5d480e6e61e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.109260] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 715.109260] env[62066]: value = "task-1155780" [ 715.109260] env[62066]: _type = "Task" [ 715.109260] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.118115] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.166309] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155776, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.225541] env[62066]: DEBUG nova.scheduler.client.report [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.311023] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.338246] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.338521] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.338681] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.338873] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.339053] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.339231] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.339450] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.339625] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.339807] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.339975] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.340211] env[62066]: DEBUG nova.virt.hardware [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.341110] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87efe64d-02ab-4464-b0d4-7de178a58b21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.351695] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e85d7df-1669-47a0-9e32-97f17875019a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.543102] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155779, 'name': CreateVM_Task, 'duration_secs': 0.358605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.545696] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 715.546387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.546387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.546669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.547424] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-507dec54-ecd2-45e8-93e1-e1592c3cc148 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.552144] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 715.552144] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52317b31-491a-37a7-61e6-a973a114c5a0" [ 715.552144] env[62066]: _type = "Task" [ 715.552144] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.564288] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52317b31-491a-37a7-61e6-a973a114c5a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.626275] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155780, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.665987] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155776, 'name': ReconfigVM_Task, 'duration_secs': 1.277367} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.666427] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf/01e1df17-4b9d-4e12-bf6b-50b39c08bfbf.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.667308] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b59b6cae-dc8d-4373-960c-9ae50ae3254f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.673534] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 715.673534] env[62066]: value = "task-1155781" [ 715.673534] env[62066]: _type = "Task" [ 715.673534] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.681025] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155781, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.730832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.731463] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.734173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.114s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.848932] env[62066]: DEBUG nova.network.neutron [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updated VIF entry in instance network info cache for port 1436b8c7-b051-48ad-962f-30ccdc95b36b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 715.849351] env[62066]: DEBUG nova.network.neutron [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.918299] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Successfully updated port: 3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.064262] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52317b31-491a-37a7-61e6-a973a114c5a0, 'name': SearchDatastore_Task, 'duration_secs': 0.027012} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.064736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.064997] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.065267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.065419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.065617] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.065875] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3936f36-2957-4764-ab1e-544406a9efc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.074720] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.074720] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 716.076235] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08864a65-e6d4-4886-8d96-eb4b0a9557ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.082664] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 716.082664] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a14c56-f0ed-0c21-5fab-cd7922dec84c" [ 716.082664] env[62066]: _type = "Task" [ 716.082664] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.090332] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a14c56-f0ed-0c21-5fab-cd7922dec84c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.119947] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155780, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610447} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.120290] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2f186d0a-91a0-4dc2-83bd-511099445af7/2f186d0a-91a0-4dc2-83bd-511099445af7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 716.120500] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.120771] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7eeb997-3291-43c0-b0f1-8f1cf0cd7226 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.127043] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 716.127043] env[62066]: value = "task-1155782" [ 716.127043] env[62066]: _type = "Task" [ 716.127043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.134668] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155782, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.183586] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155781, 'name': Rename_Task, 'duration_secs': 0.133076} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.183708] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 716.185859] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3966ff25-2e61-499e-810d-db0289a546f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.190234] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 716.190234] env[62066]: value = "task-1155783" [ 716.190234] env[62066]: _type = "Task" [ 716.190234] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.198399] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.240681] env[62066]: DEBUG nova.compute.utils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 716.247325] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.248634] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 716.292222] env[62066]: DEBUG nova.policy [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315ec256cd6b422a90f2914175cb49bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c93c0e92cfec42f4b4a20e9fb4a32088', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.352245] env[62066]: DEBUG oslo_concurrency.lockutils [req-6b531457-f948-4673-bc1b-5d2f749af956 req-56337ea7-217b-4c09-87bc-1594d70ba20b service nova] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.421742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.421843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.421967] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.437012] env[62066]: DEBUG nova.compute.manager [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Received event network-vif-plugged-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.437371] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Acquiring lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.437625] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.438683] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.438683] env[62066]: DEBUG nova.compute.manager [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] No waiting events found dispatching network-vif-plugged-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 716.438683] env[62066]: WARNING nova.compute.manager [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Received unexpected event network-vif-plugged-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 for instance with vm_state building and task_state spawning. [ 716.438683] env[62066]: DEBUG nova.compute.manager [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Received event network-changed-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.439454] env[62066]: DEBUG nova.compute.manager [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Refreshing instance network info cache due to event network-changed-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.439454] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Acquiring lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.595630] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a14c56-f0ed-0c21-5fab-cd7922dec84c, 'name': SearchDatastore_Task, 'duration_secs': 0.046056} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.596693] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3153d15-08d3-4656-b4af-643438625349 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.600901] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304bd7ca-3310-422a-8544-7fcef9ba3869 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.607379] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 716.607379] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524319c6-66d0-3e27-f07c-9cdf0ea0a59a" [ 716.607379] env[62066]: _type = "Task" [ 716.607379] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.613293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e823ee5-d655-466c-843c-77319fcdaf5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.622699] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524319c6-66d0-3e27-f07c-9cdf0ea0a59a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.653718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fa8d30-9296-4805-b227-87f7c1d6eace {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.663997] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155782, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.206553} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.664414] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 716.665685] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7eef4a-24a8-45ec-8be5-50f671e62c9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.670248] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aaa12d-bd62-4ba9-8bb8-aa516ae90b42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.686168] env[62066]: DEBUG nova.compute.provider_tree [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.707172] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 2f186d0a-91a0-4dc2-83bd-511099445af7/2f186d0a-91a0-4dc2-83bd-511099445af7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 716.708098] env[62066]: DEBUG nova.scheduler.client.report [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.711790] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Successfully created port: 983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.723518] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23f7897e-4944-49e9-9591-4648be0c5dac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.733417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 0.999s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.733998] env[62066]: ERROR nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Traceback (most recent call last): [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.driver.spawn(context, instance, image_meta, [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] vm_ref = self.build_virtual_machine(instance, [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.733998] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] for vif in network_info: [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return self._sync_wrapper(fn, *args, **kwargs) [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.wait() [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self[:] = self._gt.wait() [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return self._exit_event.wait() [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] current.throw(*self._exc) [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.734562] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] result = function(*args, **kwargs) [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] return func(*args, **kwargs) [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise e [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] nwinfo = self.network_api.allocate_for_instance( [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] created_port_ids = self._update_ports_for_instance( [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] with excutils.save_and_reraise_exception(): [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] self.force_reraise() [ 716.735131] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise self.value [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] updated_port = self._update_port( [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] _ensure_no_port_binding_failure(port) [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] raise exception.PortBindingFailed(port_id=port['id']) [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] nova.exception.PortBindingFailed: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. [ 716.735669] env[62066]: ERROR nova.compute.manager [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] [ 716.735669] env[62066]: DEBUG nova.compute.utils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 716.736112] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.023s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.737509] env[62066]: INFO nova.compute.claims [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.740338] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Build of instance ba33e43a-fbed-4a5a-9c05-a5a29082ebe9 was re-scheduled: Binding failed for port eee10f4a-57b6-40ad-8049-87544bda2f33, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.740745] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 716.740960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquiring lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.741112] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Acquired lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.741264] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.749369] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.752129] env[62066]: DEBUG oslo_vmware.api [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155783, 'name': PowerOnVM_Task, 'duration_secs': 0.504035} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.753748] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 716.753939] env[62066]: INFO nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 8.41 seconds to spawn the instance on the hypervisor. [ 716.754154] env[62066]: DEBUG nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 716.754692] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 716.754692] env[62066]: value = "task-1155784" [ 716.754692] env[62066]: _type = "Task" [ 716.754692] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.756350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cf0048-d9fa-48a2-b6c1-3417922fecaf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.770732] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.968155] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.121794] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524319c6-66d0-3e27-f07c-9cdf0ea0a59a, 'name': SearchDatastore_Task, 'duration_secs': 0.029894} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.122093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.122300] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 26d87a85-0aa3-49b9-97ca-1b7fedbebb14/26d87a85-0aa3-49b9-97ca-1b7fedbebb14.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 717.122551] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-153a8636-f962-4167-9d2d-050492e3d8f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.129912] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 717.129912] env[62066]: value = "task-1155785" [ 717.129912] env[62066]: _type = "Task" [ 717.129912] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.134887] env[62066]: DEBUG nova.network.neutron [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [{"id": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "address": "fa:16:3e:70:9f:6f", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4fe3c0-de", "ovs_interfaceid": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.138804] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.267483] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.278764] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155784, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.284067] env[62066]: INFO nova.compute.manager [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 33.25 seconds to build instance. [ 717.353897] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.640827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.641321] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Instance network_info: |[{"id": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "address": "fa:16:3e:70:9f:6f", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4fe3c0-de", "ovs_interfaceid": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.641528] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155785, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.641692] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Acquired lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.641861] env[62066]: DEBUG nova.network.neutron [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Refreshing network info cache for port 3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.643123] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:9f:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271a82f1-1d09-4ad3-9c15-07269bad114c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e4fe3c0-de42-48eb-97b5-6f00fafd3a95', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.650645] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating folder: Project (c93c0e92cfec42f4b4a20e9fb4a32088). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.654135] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b1236e8-af48-420f-85f0-a2b4a0c1f0f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.665102] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created folder: Project (c93c0e92cfec42f4b4a20e9fb4a32088) in parent group-v251573. [ 717.665192] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating folder: Instances. Parent ref: group-v251608. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.665431] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f800ef5f-5a57-4757-80d2-9f308a24838d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.676019] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created folder: Instances in parent group-v251608. [ 717.676019] env[62066]: DEBUG oslo.service.loopingcall [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.676019] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 717.676019] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a00d901f-58e3-41f7-a5c4-bd29b4359648 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.701060] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.701060] env[62066]: value = "task-1155788" [ 717.701060] env[62066]: _type = "Task" [ 717.701060] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.709392] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155788, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.757673] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.777802] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155784, 'name': ReconfigVM_Task, 'duration_secs': 0.670138} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.779950] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 2f186d0a-91a0-4dc2-83bd-511099445af7/2f186d0a-91a0-4dc2-83bd-511099445af7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 717.780891] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67260b40-cdfc-43fb-906d-624cab98836d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.786987] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.787282] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.787477] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.787732] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.787912] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.788104] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.788375] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.788581] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.788797] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.789052] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.789301] env[62066]: DEBUG nova.virt.hardware [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.789779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c1fe7557-7492-4a0e-bbde-b815a04d8019 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.596s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.790666] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0e2c60-0ec8-47cf-9404-9d628f5082c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.798537] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 717.798537] env[62066]: value = "task-1155789" [ 717.798537] env[62066]: _type = "Task" [ 717.798537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.808207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4758c3-9e82-473e-aa15-2a499b465938 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.821712] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155789, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.859394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Releasing lock "refresh_cache-ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.859639] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 717.859824] env[62066]: DEBUG nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.859987] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.877417] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.897992] env[62066]: DEBUG nova.network.neutron [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updated VIF entry in instance network info cache for port 3e4fe3c0-de42-48eb-97b5-6f00fafd3a95. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.898318] env[62066]: DEBUG nova.network.neutron [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [{"id": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "address": "fa:16:3e:70:9f:6f", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4fe3c0-de", "ovs_interfaceid": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.136465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ebe161-820f-4747-b942-3aeeebacc2da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.144507] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64082} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.146023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 26d87a85-0aa3-49b9-97ca-1b7fedbebb14/26d87a85-0aa3-49b9-97ca-1b7fedbebb14.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 718.146213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.146380] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-67213291-9af2-48c0-b73f-6dbe9496e142 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.149234] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d9ad43-25ea-4b53-8947-b0aa5badc846 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.183095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed66fce-f1c0-4b8f-99ec-01d9b36c199a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.185860] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 718.185860] env[62066]: value = "task-1155790" [ 718.185860] env[62066]: _type = "Task" [ 718.185860] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.192764] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c48893-f8bc-4ae2-99a6-c54f845d0ed1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.201124] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.212756] env[62066]: DEBUG nova.compute.provider_tree [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.220550] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155788, 'name': CreateVM_Task, 'duration_secs': 0.496905} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.220550] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.221216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.221216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.221544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.222361] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8489fe58-4664-4f8f-94d5-65a7b0b9735b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.231030] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 718.231030] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e99f46-9ba8-95cd-0fa6-d481fe1fab6d" [ 718.231030] env[62066]: _type = "Task" [ 718.231030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.236195] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e99f46-9ba8-95cd-0fa6-d481fe1fab6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.300356] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.312762] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155789, 'name': Rename_Task, 'duration_secs': 0.328049} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.312762] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 718.312762] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ad917eb-a94c-44db-8079-cca7a1b5dcf5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.317431] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Successfully updated port: 983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 718.319707] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 718.319707] env[62066]: value = "task-1155791" [ 718.319707] env[62066]: _type = "Task" [ 718.319707] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.328232] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155791, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.382214] env[62066]: DEBUG nova.network.neutron [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.401410] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc8fa5ee-f8dd-4b30-a376-e715ce2eef7c req-fcd5d014-5245-4a6a-81a6-aed49371de2e service nova] Releasing lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.469266] env[62066]: DEBUG nova.compute.manager [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Received event network-vif-plugged-983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.469560] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Acquiring lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.469687] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.469854] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.470104] env[62066]: DEBUG nova.compute.manager [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] No waiting events found dispatching network-vif-plugged-983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.470306] env[62066]: WARNING nova.compute.manager [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Received unexpected event network-vif-plugged-983e5ae8-8e7c-41f2-9304-39f8dc1b8347 for instance with vm_state building and task_state spawning. [ 718.470468] env[62066]: DEBUG nova.compute.manager [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Received event network-changed-983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.470623] env[62066]: DEBUG nova.compute.manager [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Refreshing instance network info cache due to event network-changed-983e5ae8-8e7c-41f2-9304-39f8dc1b8347. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.470805] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Acquiring lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.470938] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Acquired lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.471100] env[62066]: DEBUG nova.network.neutron [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Refreshing network info cache for port 983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.695501] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.695771] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.696536] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d77ff25-e2b0-4a28-b2d4-b722c54ca86a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.718084] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 26d87a85-0aa3-49b9-97ca-1b7fedbebb14/26d87a85-0aa3-49b9-97ca-1b7fedbebb14.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.718889] env[62066]: DEBUG nova.scheduler.client.report [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.721812] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef0a14b8-5696-45de-b89f-c0e01b04d771 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.737708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.738168] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.740817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.769s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.740999] env[62066]: DEBUG nova.objects.instance [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 718.752758] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e99f46-9ba8-95cd-0fa6-d481fe1fab6d, 'name': SearchDatastore_Task, 'duration_secs': 0.03056} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.754059] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.754380] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 718.754512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.754654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.754830] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 718.755145] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 718.755145] env[62066]: value = "task-1155792" [ 718.755145] env[62066]: _type = "Task" [ 718.755145] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.755320] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b2ac065-e863-4469-af20-29315cdea0d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.765998] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155792, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.768077] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.768153] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 718.769223] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d449aaf-23f4-4ffe-83da-33f07459c261 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.776383] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 718.776383] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52189c57-a33a-08ad-32f6-bd063bc821f5" [ 718.776383] env[62066]: _type = "Task" [ 718.776383] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.786610] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52189c57-a33a-08ad-32f6-bd063bc821f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.821529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.826287] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.831463] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155791, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.885842] env[62066]: INFO nova.compute.manager [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] [instance: ba33e43a-fbed-4a5a-9c05-a5a29082ebe9] Took 1.03 seconds to deallocate network for instance. [ 719.002945] env[62066]: DEBUG nova.network.neutron [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.074245] env[62066]: DEBUG nova.network.neutron [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.245078] env[62066]: DEBUG nova.compute.utils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 719.250742] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.250742] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.268854] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155792, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.287136] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52189c57-a33a-08ad-32f6-bd063bc821f5, 'name': SearchDatastore_Task, 'duration_secs': 0.036868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.288248] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91dc4250-e7fb-4778-b08a-7e9d485f79da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.293796] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 719.293796] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523158b0-23d6-0472-ef0c-248cdfb3e195" [ 719.293796] env[62066]: _type = "Task" [ 719.293796] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.298816] env[62066]: DEBUG nova.policy [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e36a41f0ee3a4bd5815e5067b4114ac4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00dcb6ace2714cd4994a5273586c5b95', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.304305] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523158b0-23d6-0472-ef0c-248cdfb3e195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.331566] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155791, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.577444] env[62066]: DEBUG oslo_concurrency.lockutils [req-e5401a48-76e8-4687-a6d9-c9d8db483146 req-24962701-8fed-454e-b71a-550032e81db1 service nova] Releasing lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.577694] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.577892] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.606545] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully created port: ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.751534] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.758020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ead3d9c9-7179-45b5-a37b-aa3b568204a5 tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.758020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.661s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.758020] env[62066]: INFO nova.compute.claims [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.769470] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155792, 'name': ReconfigVM_Task, 'duration_secs': 0.631981} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.770480] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 26d87a85-0aa3-49b9-97ca-1b7fedbebb14/26d87a85-0aa3-49b9-97ca-1b7fedbebb14.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.770875] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc2021be-c3c9-463d-bfcc-9d19e771e374 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.776671] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 719.776671] env[62066]: value = "task-1155793" [ 719.776671] env[62066]: _type = "Task" [ 719.776671] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.785212] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155793, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.807025] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523158b0-23d6-0472-ef0c-248cdfb3e195, 'name': SearchDatastore_Task, 'duration_secs': 0.011676} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.807025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.807025] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/d50f0505-66f4-412f-9744-25e0ea96277a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 719.807025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbd297f0-93da-49b1-9941-65f702772d75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.812468] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 719.812468] env[62066]: value = "task-1155794" [ 719.812468] env[62066]: _type = "Task" [ 719.812468] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.826299] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.833436] env[62066]: DEBUG oslo_vmware.api [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155791, 'name': PowerOnVM_Task, 'duration_secs': 1.418569} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.833914] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.834254] env[62066]: INFO nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Took 8.99 seconds to spawn the instance on the hypervisor. [ 719.834547] env[62066]: DEBUG nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.835434] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe8c755-77aa-4a24-b114-ba47cca8beaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.901248] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully created port: 5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.926716] env[62066]: INFO nova.scheduler.client.report [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Deleted allocations for instance ba33e43a-fbed-4a5a-9c05-a5a29082ebe9 [ 720.127209] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.212632] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 720.212876] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 720.247614] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully created port: 77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.295387] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155793, 'name': Rename_Task, 'duration_secs': 0.173268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.295789] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 720.296154] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce84bcd8-adf9-4004-9c36-448105e24cd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.304469] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 720.304469] env[62066]: value = "task-1155795" [ 720.304469] env[62066]: _type = "Task" [ 720.304469] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.314194] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155795, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.328031] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155794, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.355930] env[62066]: INFO nova.compute.manager [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Took 33.26 seconds to build instance. [ 720.401871] env[62066]: DEBUG nova.network.neutron [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Updating instance_info_cache with network_info: [{"id": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "address": "fa:16:3e:98:2b:0a", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap983e5ae8-8e", "ovs_interfaceid": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.434074] env[62066]: DEBUG oslo_concurrency.lockutils [None req-55d8def1-7509-46fb-8718-398170b3133b tempest-ServersTestBootFromVolume-1642038895 tempest-ServersTestBootFromVolume-1642038895-project-member] Lock "ba33e43a-fbed-4a5a-9c05-a5a29082ebe9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.105s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.719250] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 720.719708] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 720.719837] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 720.772425] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.804166] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.804406] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.804559] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.804731] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.804874] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.805102] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.805234] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.805392] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.805557] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.805715] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.805877] env[62066]: DEBUG nova.virt.hardware [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.807091] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19080877-4ea3-4f7a-acd4-67aff100294a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.827118] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155795, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.829654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c32af5-b1b0-4903-9c26-76cce4c3db51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.849773] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64111} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.850262] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/d50f0505-66f4-412f-9744-25e0ea96277a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 720.851042] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.851042] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51d9fe67-4722-4f33-a941-8a1f0361a923 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.858229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7024ce56-caca-4ab3-b028-1e92bcd427a4 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.338s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.859895] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 720.859895] env[62066]: value = "task-1155797" [ 720.859895] env[62066]: _type = "Task" [ 720.859895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.870434] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155797, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.905811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.906134] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance network_info: |[{"id": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "address": "fa:16:3e:98:2b:0a", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap983e5ae8-8e", "ovs_interfaceid": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 720.906730] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:2b:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271a82f1-1d09-4ad3-9c15-07269bad114c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '983e5ae8-8e7c-41f2-9304-39f8dc1b8347', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.914621] env[62066]: DEBUG oslo.service.loopingcall [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.915236] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.915476] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-795e0f4a-40fb-4dec-892b-e892d2c8815c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.936924] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.940862] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.940862] env[62066]: value = "task-1155798" [ 720.940862] env[62066]: _type = "Task" [ 720.940862] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.954389] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155798, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.211537] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a744b98-2854-47aa-a9c0-addbf4098d11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.219456] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb83994-daa4-4bd8-b56c-167a1872037d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.224705] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 721.224853] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 721.224979] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 721.225144] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 721.225349] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 721.257926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.258255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.258255] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 721.258433] env[62066]: DEBUG nova.objects.instance [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lazy-loading 'info_cache' on Instance uuid d6837004-9b24-45f7-9ed7-48c4503859cb {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 721.263021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d5898f-a555-45e8-a85e-0c4478d32410 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.268359] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40cb337-9529-4392-8138-d2378bb223f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.283302] env[62066]: DEBUG nova.compute.provider_tree [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.315705] env[62066]: DEBUG oslo_vmware.api [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1155795, 'name': PowerOnVM_Task, 'duration_secs': 0.517596} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.315981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 721.316215] env[62066]: INFO nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Took 8.45 seconds to spawn the instance on the hypervisor. [ 721.316393] env[62066]: DEBUG nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.317221] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cecb47-8178-4b43-94be-be9363c56f37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.367347] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.373266] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066829} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.373692] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.374530] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f71cb5-0906-4e77-95a9-3fca738d5584 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.398850] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/d50f0505-66f4-412f-9744-25e0ea96277a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.399493] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65a8bfde-3059-47d1-8b6c-030bc1701ad6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.428385] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 721.428385] env[62066]: value = "task-1155799" [ 721.428385] env[62066]: _type = "Task" [ 721.428385] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.437411] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.458305] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155798, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.465036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.789400] env[62066]: DEBUG nova.scheduler.client.report [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.836394] env[62066]: INFO nova.compute.manager [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Took 32.83 seconds to build instance. [ 721.891109] env[62066]: DEBUG nova.compute.manager [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-plugged-ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.891109] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.891109] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.891109] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.891109] env[62066]: DEBUG nova.compute.manager [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] No waiting events found dispatching network-vif-plugged-ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 721.891289] env[62066]: WARNING nova.compute.manager [req-cd9fe27f-1764-4552-9adb-619f7b33efb6 req-0d86d315-02fd-4c82-a8e4-bfcacbca19c1 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received unexpected event network-vif-plugged-ab2ecd7c-9249-477c-aeef-56105dc12410 for instance with vm_state building and task_state spawning. [ 721.899548] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.938610] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.957162] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155798, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.012457] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully updated port: ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 722.021887] env[62066]: DEBUG nova.compute.manager [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.023080] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e49ff3-28c7-462b-adb9-d4f23998b51e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.295552] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.298497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.298497] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.301421] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.237s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.301629] env[62066]: DEBUG nova.objects.instance [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lazy-loading 'resources' on Instance uuid d6837004-9b24-45f7-9ed7-48c4503859cb {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 722.339454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8004661c-7794-49b4-830f-d55ccba76267 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.489s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.438686] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155799, 'name': ReconfigVM_Task, 'duration_secs': 0.607008} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.439024] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfigured VM instance instance-0000002d to attach disk [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/d50f0505-66f4-412f-9744-25e0ea96277a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.440087] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d444f58-bc5b-43f5-b243-a2b06582b659 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.447051] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 722.447051] env[62066]: value = "task-1155800" [ 722.447051] env[62066]: _type = "Task" [ 722.447051] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.458641] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155798, 'name': CreateVM_Task, 'duration_secs': 1.311331} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.461912] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 722.462257] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155800, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.462995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.463208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.463430] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 722.463689] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc96faa2-515f-4db2-b39e-3b6a2bfd192f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.468276] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 722.468276] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5203f511-c4a3-bbf3-cff1-1d072a755487" [ 722.468276] env[62066]: _type = "Task" [ 722.468276] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.478441] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5203f511-c4a3-bbf3-cff1-1d072a755487, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.537421] env[62066]: INFO nova.compute.manager [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] instance snapshotting [ 722.540601] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82587755-4552-4f04-b56e-f6aee362f5dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.562087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10363a27-6e2f-4ada-bb3a-307fad79425f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.805210] env[62066]: DEBUG nova.compute.utils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 722.809220] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.809409] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.843725] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.897188] env[62066]: DEBUG nova.policy [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c41065776f39438a8f6d427c0f62b93f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c468d5ba348d437f97a74e0da70bb42e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.933970] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.963564] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155800, 'name': Rename_Task, 'duration_secs': 0.138011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.963981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 722.964308] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3a9865f-c6db-4aaa-a69b-bfa3f6f20dd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.976209] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 722.976209] env[62066]: value = "task-1155801" [ 722.976209] env[62066]: _type = "Task" [ 722.976209] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.985051] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5203f511-c4a3-bbf3-cff1-1d072a755487, 'name': SearchDatastore_Task, 'duration_secs': 0.010894} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.988124] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.988395] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 722.988824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.989100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.989407] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.993971] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7090610f-bc92-4003-a49f-ae9bec004914 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.995035] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155801, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.007105] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.007358] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.008166] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d4f4472-f9e7-498c-97ea-d571d5cf6c67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.014432] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 723.014432] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dd2a2c-ea0b-2a75-8be4-8f6ad2ea4845" [ 723.014432] env[62066]: _type = "Task" [ 723.014432] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.036839] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52dd2a2c-ea0b-2a75-8be4-8f6ad2ea4845, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.091096] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 723.091971] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-37e1691c-2d3f-44f8-95c3-88199d8d6753 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.100340] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 723.100340] env[62066]: value = "task-1155802" [ 723.100340] env[62066]: _type = "Task" [ 723.100340] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.115274] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155802, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.275893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c11811-4367-4de0-9886-e59e73cc17a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.285738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe055ce-f66c-4392-996b-e464974c76cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.325338] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.329886] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dd8980-bf23-456d-ab06-6395ece7df65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.338372] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15fd437-e41b-43f8-8e88-12805d9a4191 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.361180] env[62066]: DEBUG nova.compute.provider_tree [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.383021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.436367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-d6837004-9b24-45f7-9ed7-48c4503859cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.436592] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 723.436819] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.436997] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.437182] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.437325] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.437478] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.437626] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.437755] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 723.437910] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 723.489543] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155801, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.529250] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52dd2a2c-ea0b-2a75-8be4-8f6ad2ea4845, 'name': SearchDatastore_Task, 'duration_secs': 0.018205} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.529250] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6517b5c4-9b7c-4a78-bb2b-b818dac20d01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.535205] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 723.535205] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527881c4-bc18-ddda-b9da-93ee8fb14cc1" [ 723.535205] env[62066]: _type = "Task" [ 723.535205] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.545483] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527881c4-bc18-ddda-b9da-93ee8fb14cc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.610773] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155802, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.730805] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Successfully created port: 81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.865034] env[62066]: DEBUG nova.scheduler.client.report [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.941922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.990763] env[62066]: DEBUG oslo_vmware.api [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155801, 'name': PowerOnVM_Task, 'duration_secs': 0.563957} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.991205] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 723.991324] env[62066]: INFO nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Took 8.68 seconds to spawn the instance on the hypervisor. [ 723.994095] env[62066]: DEBUG nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.994944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307103f6-77eb-4e6a-8870-0a8c7357b2d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.046033] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527881c4-bc18-ddda-b9da-93ee8fb14cc1, 'name': SearchDatastore_Task, 'duration_secs': 0.009202} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.046364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.046702] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/a8dd7483-0588-4f60-9504-20de799e69f1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.046961] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9196ffc-b521-44b1-a012-590abe8e2ae5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.054688] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 724.054688] env[62066]: value = "task-1155803" [ 724.054688] env[62066]: _type = "Task" [ 724.054688] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.064916] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.111796] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155802, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.195931] env[62066]: DEBUG nova.compute.manager [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-changed-ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.196190] env[62066]: DEBUG nova.compute.manager [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing instance network info cache due to event network-changed-ab2ecd7c-9249-477c-aeef-56105dc12410. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.196481] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Acquiring lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.196705] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Acquired lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.196882] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing network info cache for port ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.337356] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.370902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.069s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.373012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.779s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.374811] env[62066]: INFO nova.compute.claims [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.381414] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.381707] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.381900] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.382365] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.382540] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.382686] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.382888] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.383077] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.383303] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.383511] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.383761] env[62066]: DEBUG nova.virt.hardware [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.384956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edac306d-0196-4ced-b2de-37383b451ff4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.395040] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72759dd7-5225-4b8c-84ce-b9f64d25e8e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.414955] env[62066]: INFO nova.scheduler.client.report [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Deleted allocations for instance d6837004-9b24-45f7-9ed7-48c4503859cb [ 724.493344] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully updated port: 5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 724.513829] env[62066]: INFO nova.compute.manager [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Took 29.20 seconds to build instance. [ 724.566059] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155803, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.614054] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155802, 'name': CreateSnapshot_Task, 'duration_secs': 1.029902} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.614367] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 724.615172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da185db-8da0-4077-9185-aff1eecc3c1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.745225] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.895741] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.927353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0710dc3-b039-409d-ae19-1e610b00e24b tempest-ServerShowV257Test-1406587039 tempest-ServerShowV257Test-1406587039-project-member] Lock "d6837004-9b24-45f7-9ed7-48c4503859cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.603s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.015966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36d7a75f-aba0-4ade-8468-028043d94a0f tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.783s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.070251] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.836265} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.070251] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/a8dd7483-0588-4f60-9504-20de799e69f1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 725.070251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 725.070251] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e89e3ae2-3a2a-454e-a14e-6111f1dc71f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.077710] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 725.077710] env[62066]: value = "task-1155804" [ 725.077710] env[62066]: _type = "Task" [ 725.077710] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.092229] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155804, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.137713] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 725.138425] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d59c6e8c-fb65-4628-b06e-4ff93753956d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.149585] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 725.149585] env[62066]: value = "task-1155805" [ 725.149585] env[62066]: _type = "Task" [ 725.149585] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.158131] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.397976] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Releasing lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.398324] env[62066]: DEBUG nova.compute.manager [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-changed-1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.398431] env[62066]: DEBUG nova.compute.manager [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing instance network info cache due to event network-changed-1436b8c7-b051-48ad-962f-30ccdc95b36b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.398653] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.398792] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.398954] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing network info cache for port 1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.520421] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.601401] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155804, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07508} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.602680] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.603159] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1713b9e0-dfdf-4e8e-a164-7999780b8bca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.626372] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/a8dd7483-0588-4f60-9504-20de799e69f1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.629430] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef63e7bb-4bf2-405e-9ff7-b94ac12967bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.670279] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.675669] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 725.675669] env[62066]: value = "task-1155806" [ 725.675669] env[62066]: _type = "Task" [ 725.675669] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.691404] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155806, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.870747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41a72e8-58aa-4600-bba5-b3319387cac7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.878398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-552051c6-f7e5-4e97-a71c-001a43479c91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.915679] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb66475-3968-43c5-a7b4-e012b81b20f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.924917] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd35e1c-b4fe-49a0-9736-3374d8110360 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.941453] env[62066]: DEBUG nova.compute.provider_tree [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.049128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.075144] env[62066]: DEBUG nova.compute.manager [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Received event network-vif-plugged-81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.075144] env[62066]: DEBUG oslo_concurrency.lockutils [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.075144] env[62066]: DEBUG oslo_concurrency.lockutils [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.075144] env[62066]: DEBUG oslo_concurrency.lockutils [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.075144] env[62066]: DEBUG nova.compute.manager [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] No waiting events found dispatching network-vif-plugged-81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 726.076314] env[62066]: WARNING nova.compute.manager [req-dd75a9fb-0380-4ab0-a25f-f0be33c54d56 req-53c4295c-ff1f-44b1-8b87-5f0839bc1cdc service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Received unexpected event network-vif-plugged-81a78519-507f-4de8-8e08-3a469f901c85 for instance with vm_state building and task_state spawning. [ 726.076314] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Successfully updated port: 81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 726.171762] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.192937] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155806, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.286119] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updated VIF entry in instance network info cache for port 1436b8c7-b051-48ad-962f-30ccdc95b36b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.286119] env[62066]: DEBUG nova.network.neutron [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.449416] env[62066]: DEBUG nova.scheduler.client.report [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.454222] env[62066]: DEBUG nova.compute.manager [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-plugged-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.454222] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.454222] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.454222] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.454222] env[62066]: DEBUG nova.compute.manager [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] No waiting events found dispatching network-vif-plugged-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 726.454529] env[62066]: WARNING nova.compute.manager [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received unexpected event network-vif-plugged-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 for instance with vm_state building and task_state spawning. [ 726.454529] env[62066]: DEBUG nova.compute.manager [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-changed-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.454529] env[62066]: DEBUG nova.compute.manager [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing instance network info cache due to event network-changed-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.454529] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Acquiring lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.454529] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Acquired lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.454689] env[62066]: DEBUG nova.network.neutron [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing network info cache for port 5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.577507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.577625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.577843] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.598898] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Successfully updated port: 77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 726.670479] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.687994] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155806, 'name': ReconfigVM_Task, 'duration_secs': 0.739504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.688359] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Reconfigured VM instance instance-0000002e to attach disk [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/a8dd7483-0588-4f60-9504-20de799e69f1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.689018] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f768818-ab36-43ed-9d73-f778e9cf90ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.696082] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 726.696082] env[62066]: value = "task-1155807" [ 726.696082] env[62066]: _type = "Task" [ 726.696082] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.704913] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155807, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.790902] env[62066]: DEBUG oslo_concurrency.lockutils [req-da3eaf5a-3cdc-4b26-9b21-17867a6c685c req-637c489c-207b-4404-9be9-938b900c47ad service nova] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.956828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.957382] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.962426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.745s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.009996] env[62066]: DEBUG nova.network.neutron [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.102933] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.151503] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.156059] env[62066]: DEBUG nova.network.neutron [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.167629] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.207245] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155807, 'name': Rename_Task, 'duration_secs': 0.335505} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.210183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.210480] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f06d7ccd-5881-479c-b3b9-8e1f0980de4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.217823] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 727.217823] env[62066]: value = "task-1155808" [ 727.217823] env[62066]: _type = "Task" [ 727.217823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.226323] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.316933] env[62066]: DEBUG nova.network.neutron [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.464359] env[62066]: DEBUG nova.compute.utils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 727.468728] env[62066]: INFO nova.compute.claims [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.472760] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.472930] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.516315] env[62066]: DEBUG nova.policy [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85a0637645c443a9d9e6028053c0450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '350ba3c5676a4dd0a018900e7237a5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.658464] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9fc3521-b324-452c-b9f5-2f01aee06c06 req-4cdde870-2e8f-4d2d-ad94-bd5d2f8bf9e0 service nova] Releasing lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.661025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.661025] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.671652] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155805, 'name': CloneVM_Task, 'duration_secs': 2.020216} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.672557] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Created linked-clone VM from snapshot [ 727.673377] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d715a78e-32c5-4f6e-9422-bad785a9f049 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.682159] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Uploading image 218d1044-e6e2-4a2e-b8ab-4d2ff641ed5b {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 727.704534] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 727.704534] env[62066]: value = "vm-251613" [ 727.704534] env[62066]: _type = "VirtualMachine" [ 727.704534] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 727.704888] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9d963d83-003c-4304-9c6b-10a5aa96f262 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.712597] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease: (returnval){ [ 727.712597] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5263ed6c-0f65-5b86-065e-b4d4aebb9a54" [ 727.712597] env[62066]: _type = "HttpNfcLease" [ 727.712597] env[62066]: } obtained for exporting VM: (result){ [ 727.712597] env[62066]: value = "vm-251613" [ 727.712597] env[62066]: _type = "VirtualMachine" [ 727.712597] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 727.712981] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the lease: (returnval){ [ 727.712981] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5263ed6c-0f65-5b86-065e-b4d4aebb9a54" [ 727.712981] env[62066]: _type = "HttpNfcLease" [ 727.712981] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 727.720773] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 727.720773] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5263ed6c-0f65-5b86-065e-b4d4aebb9a54" [ 727.720773] env[62066]: _type = "HttpNfcLease" [ 727.720773] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 727.734636] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155808, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.785297] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Successfully created port: d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.820332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.820646] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Instance network_info: |[{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 727.821098] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:3c:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81a78519-507f-4de8-8e08-3a469f901c85', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.830211] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating folder: Project (c468d5ba348d437f97a74e0da70bb42e). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.830506] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9633f48f-fd3e-4ed9-b317-7ca76f5440db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.842167] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created folder: Project (c468d5ba348d437f97a74e0da70bb42e) in parent group-v251573. [ 727.844277] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating folder: Instances. Parent ref: group-v251614. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.844277] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f2248aa-f5a6-4618-a3cc-8986807d8fa5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.852422] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created folder: Instances in parent group-v251614. [ 727.852705] env[62066]: DEBUG oslo.service.loopingcall [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.852859] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.853074] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-244b94b3-5e0d-4ad7-8d60-f66de8e1bde0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.872884] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.872884] env[62066]: value = "task-1155812" [ 727.872884] env[62066]: _type = "Task" [ 727.872884] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.880807] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155812, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.976645] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.981385] env[62066]: INFO nova.compute.resource_tracker [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating resource usage from migration de64f433-6637-4e88-8678-c58a6683456c [ 728.173303] env[62066]: DEBUG nova.compute.manager [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Received event network-changed-81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.173375] env[62066]: DEBUG nova.compute.manager [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Refreshing instance network info cache due to event network-changed-81a78519-507f-4de8-8e08-3a469f901c85. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.173585] env[62066]: DEBUG oslo_concurrency.lockutils [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.173727] env[62066]: DEBUG oslo_concurrency.lockutils [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.173883] env[62066]: DEBUG nova.network.neutron [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Refreshing network info cache for port 81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 728.224273] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 728.224273] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5263ed6c-0f65-5b86-065e-b4d4aebb9a54" [ 728.224273] env[62066]: _type = "HttpNfcLease" [ 728.224273] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 728.224273] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.229132] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 728.229132] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5263ed6c-0f65-5b86-065e-b4d4aebb9a54" [ 728.229132] env[62066]: _type = "HttpNfcLease" [ 728.229132] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 728.233123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c022c0-d441-4b22-88e3-45a300856d32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.243029] env[62066]: DEBUG oslo_vmware.api [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155808, 'name': PowerOnVM_Task, 'duration_secs': 0.746018} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.247074] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 728.247370] env[62066]: INFO nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Took 10.49 seconds to spawn the instance on the hypervisor. [ 728.247619] env[62066]: DEBUG nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.247968] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 728.248181] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 728.256526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d81be2-4943-4823-bc0c-d3b3160df71c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.373978] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1a00ac16-7944-44e8-b562-095e376b1738 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.395080] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155812, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.508471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8cd33c-f9cf-4184-bea5-b6cb525bdf70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.521285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f855f608-c90e-4923-81ea-ba7266bbf73b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.562850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b2a795-b090-4b6f-bade-0e20cbbfb943 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.571886] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b31ce0-3104-476b-b45f-95bc8048b649 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.590391] env[62066]: DEBUG nova.compute.provider_tree [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.626407] env[62066]: DEBUG nova.compute.manager [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-plugged-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.626407] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.626637] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.626933] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.627206] env[62066]: DEBUG nova.compute.manager [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] No waiting events found dispatching network-vif-plugged-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 728.627450] env[62066]: WARNING nova.compute.manager [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received unexpected event network-vif-plugged-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 for instance with vm_state building and task_state spawning. [ 728.627694] env[62066]: DEBUG nova.compute.manager [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-changed-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.627902] env[62066]: DEBUG nova.compute.manager [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing instance network info cache due to event network-changed-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.628314] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Acquiring lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.843326] env[62066]: INFO nova.compute.manager [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Took 32.30 seconds to build instance. [ 728.868282] env[62066]: DEBUG nova.network.neutron [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [{"id": "ab2ecd7c-9249-477c-aeef-56105dc12410", "address": "fa:16:3e:86:4d:f1", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2ecd7c-92", "ovs_interfaceid": "ab2ecd7c-9249-477c-aeef-56105dc12410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "address": "fa:16:3e:b6:32:d9", "network": {"id": "f05688d3-a5c8-4a5a-85cc-ce8d7dcde861", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1381993909", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.168", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5729cb45-fe", "ovs_interfaceid": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "address": "fa:16:3e:ef:f7:2a", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77b2e9c0-f2", "ovs_interfaceid": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.885305] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155812, 'name': CreateVM_Task, 'duration_secs': 0.521607} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.885695] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 728.886620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.886957] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.887593] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 728.887887] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a7b76ea-f018-4780-b707-38f0a13ef18c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.896578] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 728.896578] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c4ebd2-6348-04bc-c1ff-9d211366ccf6" [ 728.896578] env[62066]: _type = "Task" [ 728.896578] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.908118] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c4ebd2-6348-04bc-c1ff-9d211366ccf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.996869] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.033889] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.034666] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.034666] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.034789] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.035089] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.035311] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.035608] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.035838] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.036097] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.036325] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.036555] env[62066]: DEBUG nova.virt.hardware [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.038149] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c14af1d-ee96-45f7-bd87-cd82671033b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.051465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e110ba00-9773-42d1-9702-d5882bb4a579 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.095228] env[62066]: DEBUG nova.scheduler.client.report [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.345671] env[62066]: DEBUG nova.network.neutron [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updated VIF entry in instance network info cache for port 81a78519-507f-4de8-8e08-3a469f901c85. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 729.346894] env[62066]: DEBUG nova.network.neutron [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.350018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a830461-6539-4fb5-9cd9-843a6c45b269 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.788s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.358423] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Successfully updated port: d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 729.372099] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.372515] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance network_info: |[{"id": "ab2ecd7c-9249-477c-aeef-56105dc12410", "address": "fa:16:3e:86:4d:f1", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2ecd7c-92", "ovs_interfaceid": "ab2ecd7c-9249-477c-aeef-56105dc12410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "address": "fa:16:3e:b6:32:d9", "network": {"id": "f05688d3-a5c8-4a5a-85cc-ce8d7dcde861", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1381993909", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.168", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5729cb45-fe", "ovs_interfaceid": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "address": "fa:16:3e:ef:f7:2a", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77b2e9c0-f2", "ovs_interfaceid": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 729.373149] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Acquired lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.373284] env[62066]: DEBUG nova.network.neutron [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Refreshing network info cache for port 77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.375400] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:4d:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39a4aca0-934b-4a91-8779-6a4360c3f967', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab2ecd7c-9249-477c-aeef-56105dc12410', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:32:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5729cb45-fe23-4d92-b6a2-8b5f9164d4e7', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:f7:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39a4aca0-934b-4a91-8779-6a4360c3f967', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.389519] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Creating folder: Project (00dcb6ace2714cd4994a5273586c5b95). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 729.391605] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d7a6a6f-e4b5-4f36-bcc3-398a81657f9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.403590] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Created folder: Project (00dcb6ace2714cd4994a5273586c5b95) in parent group-v251573. [ 729.403811] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Creating folder: Instances. Parent ref: group-v251617. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 729.407494] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00e93a81-1869-415f-8573-7e9b7a0ab703 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.409463] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c4ebd2-6348-04bc-c1ff-9d211366ccf6, 'name': SearchDatastore_Task, 'duration_secs': 0.01138} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.410030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.410305] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.410561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.410711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.410889] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.411516] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02dd639f-8f96-4feb-8cda-fd51aa311eef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.417599] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Created folder: Instances in parent group-v251617. [ 729.417979] env[62066]: DEBUG oslo.service.loopingcall [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.418077] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 729.418319] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee14d12c-284a-492f-aa48-52a77a7bd74d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.442020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.442020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 729.442020] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed278ae7-b74e-4449-967a-b914f75251b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.443994] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.443994] env[62066]: value = "task-1155815" [ 729.443994] env[62066]: _type = "Task" [ 729.443994] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.448734] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 729.448734] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5210fae2-7d05-f9e5-0637-dad1ce3bddef" [ 729.448734] env[62066]: _type = "Task" [ 729.448734] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.456035] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155815, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.460988] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5210fae2-7d05-f9e5-0637-dad1ce3bddef, 'name': SearchDatastore_Task, 'duration_secs': 0.008969} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.462025] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6444364f-6fd0-49d9-9e63-b9f91aed21a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.466734] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 729.466734] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523041ed-a0c6-b0d4-7cd8-a58e6e379926" [ 729.466734] env[62066]: _type = "Task" [ 729.466734] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.474912] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523041ed-a0c6-b0d4-7cd8-a58e6e379926, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.602357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.640s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.602664] env[62066]: INFO nova.compute.manager [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Migrating [ 729.603367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.603367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.604723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.151s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.607232] env[62066]: INFO nova.compute.claims [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.851320] env[62066]: DEBUG oslo_concurrency.lockutils [req-de1052f3-bb11-4e32-a711-2417d2928856 req-5248d072-5001-4a5d-a138-e4f316719d68 service nova] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.852293] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.869529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.869697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.869952] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.958322] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155815, 'name': CreateVM_Task, 'duration_secs': 0.44732} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.958322] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 729.959792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.959792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.960250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.962017] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baf508ad-716d-48cd-9dce-aa6e582b78d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.965151] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 729.965151] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529abc79-9012-6100-aa3d-d94167c0096e" [ 729.965151] env[62066]: _type = "Task" [ 729.965151] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.976976] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529abc79-9012-6100-aa3d-d94167c0096e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.982642] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523041ed-a0c6-b0d4-7cd8-a58e6e379926, 'name': SearchDatastore_Task, 'duration_secs': 0.012436} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.983262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.984026] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.984609] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fb8c55d-e0fa-4175-b14d-2322d2b5db3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.992785] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 729.992785] env[62066]: value = "task-1155816" [ 729.992785] env[62066]: _type = "Task" [ 729.992785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.001826] env[62066]: INFO nova.compute.manager [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Rescuing [ 730.002345] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.002653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.003693] env[62066]: DEBUG nova.network.neutron [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.009527] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.112338] env[62066]: INFO nova.compute.rpcapi [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 730.113141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.184071] env[62066]: DEBUG nova.network.neutron [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updated VIF entry in instance network info cache for port 77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 730.184938] env[62066]: DEBUG nova.network.neutron [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [{"id": "ab2ecd7c-9249-477c-aeef-56105dc12410", "address": "fa:16:3e:86:4d:f1", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2ecd7c-92", "ovs_interfaceid": "ab2ecd7c-9249-477c-aeef-56105dc12410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "address": "fa:16:3e:b6:32:d9", "network": {"id": "f05688d3-a5c8-4a5a-85cc-ce8d7dcde861", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1381993909", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.168", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5729cb45-fe", "ovs_interfaceid": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "address": "fa:16:3e:ef:f7:2a", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77b2e9c0-f2", "ovs_interfaceid": "77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.386193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.415449] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.478138] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529abc79-9012-6100-aa3d-d94167c0096e, 'name': SearchDatastore_Task, 'duration_secs': 0.022573} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.481692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.481692] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.481692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.481692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.481692] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.481692] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-203681dc-4b2b-4eb2-ab9c-19d866e04a22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.490881] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.491120] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.492036] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b42cf3c-0272-40f2-a957-0243be8158bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.507820] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 730.507820] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522339fd-3a7f-d34f-d031-7841a9169adc" [ 730.507820] env[62066]: _type = "Task" [ 730.507820] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.522198] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155816, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.527131] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522339fd-3a7f-d34f-d031-7841a9169adc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.598308] env[62066]: DEBUG nova.network.neutron [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Updating instance_info_cache with network_info: [{"id": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "address": "fa:16:3e:51:16:35", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd18922bc-f1", "ovs_interfaceid": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.647272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.647567] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.650969] env[62066]: DEBUG nova.network.neutron [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.693115] env[62066]: DEBUG oslo_concurrency.lockutils [req-e9ea9c8d-8134-4780-9e23-7c10ce19d182 req-e40125d3-657c-4901-bbbc-f22396614db9 service nova] Releasing lock "refresh_cache-265906be-704e-4f0f-ae33-e89135e27268" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.842687] env[62066]: DEBUG nova.compute.manager [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Received event network-vif-plugged-d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.842811] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Acquiring lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.843047] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.843250] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.843458] env[62066]: DEBUG nova.compute.manager [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] No waiting events found dispatching network-vif-plugged-d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 730.843707] env[62066]: WARNING nova.compute.manager [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Received unexpected event network-vif-plugged-d18922bc-f1a1-4553-8a99-0ec96b254d2d for instance with vm_state building and task_state spawning. [ 730.843938] env[62066]: DEBUG nova.compute.manager [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Received event network-changed-d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.844149] env[62066]: DEBUG nova.compute.manager [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Refreshing instance network info cache due to event network-changed-d18922bc-f1a1-4553-8a99-0ec96b254d2d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.844405] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Acquiring lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.010139] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155816, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561764} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.010490] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.010741] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.011025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5a6614a-7e60-4ab3-8ca9-0a534bb15851 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.026151] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522339fd-3a7f-d34f-d031-7841a9169adc, 'name': SearchDatastore_Task, 'duration_secs': 0.036664} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.028335] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 731.028335] env[62066]: value = "task-1155817" [ 731.028335] env[62066]: _type = "Task" [ 731.028335] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.028541] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bda184b-f568-4041-b961-9a57c66716ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.040759] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155817, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.044455] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 731.044455] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2515b-7170-2509-57f5-956c95d228c7" [ 731.044455] env[62066]: _type = "Task" [ 731.044455] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.053187] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2515b-7170-2509-57f5-956c95d228c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.091590] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3ff758-fff0-4307-970d-5ad266757e5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.097763] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c61cbf3-6d7e-4084-ada1-0d8956c5d7ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.101599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.102135] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance network_info: |[{"id": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "address": "fa:16:3e:51:16:35", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd18922bc-f1", "ovs_interfaceid": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 731.102276] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Acquired lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.102747] env[62066]: DEBUG nova.network.neutron [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Refreshing network info cache for port d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 731.103713] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:16:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd18922bc-f1a1-4553-8a99-0ec96b254d2d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 731.111672] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating folder: Project (350ba3c5676a4dd0a018900e7237a5a5). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 731.113683] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-054b2b6d-eef2-44ef-8949-5747dde47d8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.143793] env[62066]: DEBUG nova.network.neutron [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Updating instance_info_cache with network_info: [{"id": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "address": "fa:16:3e:98:2b:0a", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap983e5ae8-8e", "ovs_interfaceid": "983e5ae8-8e7c-41f2-9304-39f8dc1b8347", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.146238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8950a3b-a6da-417f-b183-93d383ec2d1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.161308] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34b36f2-b643-4ef3-b6ed-c43e0614619e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.166493] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created folder: Project (350ba3c5676a4dd0a018900e7237a5a5) in parent group-v251573. [ 731.166722] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating folder: Instances. Parent ref: group-v251620. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 731.167393] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d29ba081-5cb2-459e-afdd-f456a73dea02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.184328] env[62066]: DEBUG nova.compute.provider_tree [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.187126] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created folder: Instances in parent group-v251620. [ 731.187512] env[62066]: DEBUG oslo.service.loopingcall [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.187827] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 731.188050] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8142339f-87cb-40a3-8a54-5a91b5ace5d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.210033] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.210033] env[62066]: value = "task-1155820" [ 731.210033] env[62066]: _type = "Task" [ 731.210033] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.221386] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155820, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.262871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "4723be94-3479-4e66-8088-914824c0e669" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.264056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.546103] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155817, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070379} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.546103] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.546103] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e96e03-c952-4e78-8867-a985b227eb85 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.557402] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2515b-7170-2509-57f5-956c95d228c7, 'name': SearchDatastore_Task, 'duration_secs': 0.014174} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.567676] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.568038] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 265906be-704e-4f0f-ae33-e89135e27268/265906be-704e-4f0f-ae33-e89135e27268.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 731.577435] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.577795] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1a4eb11-67f1-42fa-916b-069248d0212a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.580118] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44536c2b-6bbc-4118-806b-ca639ddb97d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.596399] env[62066]: DEBUG nova.network.neutron [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.603445] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 731.603445] env[62066]: value = "task-1155821" [ 731.603445] env[62066]: _type = "Task" [ 731.603445] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.605296] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 731.605296] env[62066]: value = "task-1155822" [ 731.605296] env[62066]: _type = "Task" [ 731.605296] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.620044] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.624524] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.650396] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-a8dd7483-0588-4f60-9504-20de799e69f1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.712107] env[62066]: ERROR nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [req-de1cf152-40fb-4a8d-b93d-e61f9984b942] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-de1cf152-40fb-4a8d-b93d-e61f9984b942"}]} [ 731.726173] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155820, 'name': CreateVM_Task, 'duration_secs': 0.456003} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.726364] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 731.727227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.727415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.728652] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 731.728652] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51da8f2e-e369-4b7b-a4fd-cc3544ffa68f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.738022] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 731.738022] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e20a16-3c65-9a48-fb03-fe122990b730" [ 731.738022] env[62066]: _type = "Task" [ 731.738022] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.740161] env[62066]: DEBUG nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 731.748835] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e20a16-3c65-9a48-fb03-fe122990b730, 'name': SearchDatastore_Task, 'duration_secs': 0.00877} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.748835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.748835] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.748835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.748835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.748835] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.749265] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b5b2281-d274-44d2-9ff2-677a7e1a54b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.762823] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.763063] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 731.766356] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b25e667-b45d-44c4-a07a-6d48868f8ed3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.772102] env[62066]: DEBUG nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 731.772344] env[62066]: DEBUG nova.compute.provider_tree [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.779872] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 731.779872] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bafb4e-c385-a698-ebe8-657af57f0b46" [ 731.779872] env[62066]: _type = "Task" [ 731.779872] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.790043] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bafb4e-c385-a698-ebe8-657af57f0b46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.792130] env[62066]: DEBUG nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 731.825986] env[62066]: DEBUG nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 732.084396] env[62066]: DEBUG nova.network.neutron [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Updated VIF entry in instance network info cache for port d18922bc-f1a1-4553-8a99-0ec96b254d2d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 732.084923] env[62066]: DEBUG nova.network.neutron [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Updating instance_info_cache with network_info: [{"id": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "address": "fa:16:3e:51:16:35", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd18922bc-f1", "ovs_interfaceid": "d18922bc-f1a1-4553-8a99-0ec96b254d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.100070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.116311] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155821, 'name': ReconfigVM_Task, 'duration_secs': 0.331066} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.119578] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfigured VM instance instance-00000030 to attach disk [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.121479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7045ac0-3e25-471b-b8f1-a5e67cb3757d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.125334] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486125} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.125930] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 265906be-704e-4f0f-ae33-e89135e27268/265906be-704e-4f0f-ae33-e89135e27268.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.126162] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.126394] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7632e3d7-fba5-4627-8033-b4085de8f031 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.130451] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 732.130451] env[62066]: value = "task-1155823" [ 732.130451] env[62066]: _type = "Task" [ 732.130451] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.136632] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 732.136632] env[62066]: value = "task-1155824" [ 732.136632] env[62066]: _type = "Task" [ 732.136632] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.143896] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155823, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.148299] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155824, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.192658] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 732.193583] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-375624fb-27ed-49f7-af5e-c428e4a1e3f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.200459] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 732.200459] env[62066]: value = "task-1155825" [ 732.200459] env[62066]: _type = "Task" [ 732.200459] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.212513] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.255764] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ab65ac-2c8d-46a9-bfab-cb00cf4e915a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.264216] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5d5b4c-246c-469d-83a6-fd96294ab878 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.300068] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f29fea-a379-4269-89da-a1f7249a2834 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.309835] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bafb4e-c385-a698-ebe8-657af57f0b46, 'name': SearchDatastore_Task, 'duration_secs': 0.060548} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.312633] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-981ec458-daf7-43f9-a724-43620721e69c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.315934] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5c80f9-95fa-46f7-86f2-0d668dc04595 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.322772] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 732.322772] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521f3c3b-299d-4709-c888-6a16e76d5262" [ 732.322772] env[62066]: _type = "Task" [ 732.322772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.331130] env[62066]: DEBUG nova.compute.provider_tree [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.341025] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521f3c3b-299d-4709-c888-6a16e76d5262, 'name': SearchDatastore_Task, 'duration_secs': 0.009141} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.342128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.342417] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 732.342693] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0d97a1e-b05c-4e84-9522-5a22f006386f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.350718] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 732.350718] env[62066]: value = "task-1155826" [ 732.350718] env[62066]: _type = "Task" [ 732.350718] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.363338] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155826, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.588558] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa9833b8-28fb-420d-aabe-b8f03d96b824 req-4e09f409-980a-4568-ad00-0d72b41574a2 service nova] Releasing lock "refresh_cache-6a498481-b2ca-4813-87b7-2f09dfa107f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.645992] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155823, 'name': Rename_Task, 'duration_secs': 0.147317} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.646963] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 732.647337] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75e331fe-8f0c-42b1-b903-be4d65682163 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.653653] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155824, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06075} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.654981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.655594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cd96f8-be2b-40cc-b006-165dea9af2b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.660586] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 732.660586] env[62066]: value = "task-1155827" [ 732.660586] env[62066]: _type = "Task" [ 732.660586] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.688163] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 265906be-704e-4f0f-ae33-e89135e27268/265906be-704e-4f0f-ae33-e89135e27268.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.689155] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-581771fd-f0b4-4a5b-b5c1-b22fbb38c09d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.710825] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155827, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.720540] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155825, 'name': PowerOffVM_Task, 'duration_secs': 0.278499} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.722144] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.722627] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 732.722627] env[62066]: value = "task-1155828" [ 732.722627] env[62066]: _type = "Task" [ 732.722627] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.723285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ba47c1-95ca-4fa4-8d85-d79748cc28f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.736053] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.752103] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90503518-fb34-44fb-8206-e16e56d68ce7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.786058] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 732.786922] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78149a68-bfa5-4725-abc6-db3d0831fd70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.793621] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 732.793621] env[62066]: value = "task-1155829" [ 732.793621] env[62066]: _type = "Task" [ 732.793621] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.804081] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 732.804535] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.804846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.805346] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.805741] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.806905] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47f8d18b-eeb5-4278-813e-0ff67ebc3b27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.817508] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.817735] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 732.818540] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dba9b92-1a3b-4401-9b3d-00d1df8f2ebe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.824216] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 732.824216] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525e652d-23af-7a8c-dcde-c562245a415f" [ 732.824216] env[62066]: _type = "Task" [ 732.824216] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.832406] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525e652d-23af-7a8c-dcde-c562245a415f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.860800] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155826, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.861117] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.861432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.861803] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-371b3924-72fe-463c-9334-99ddbbaf595b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.866615] env[62066]: DEBUG nova.scheduler.client.report [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 74 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 732.866928] env[62066]: DEBUG nova.compute.provider_tree [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 74 to 75 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 732.867154] env[62066]: DEBUG nova.compute.provider_tree [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.871578] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 732.871578] env[62066]: value = "task-1155830" [ 732.871578] env[62066]: _type = "Task" [ 732.871578] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.880413] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155830, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.173015] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155827, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.239042] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.334797] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525e652d-23af-7a8c-dcde-c562245a415f, 'name': SearchDatastore_Task, 'duration_secs': 0.028472} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.335999] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f869b9d-b155-48dd-9ea3-28006f451a72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.340893] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 733.340893] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c94ca9-b537-4a48-70a0-b6eb8e1393c4" [ 733.340893] env[62066]: _type = "Task" [ 733.340893] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.349125] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c94ca9-b537-4a48-70a0-b6eb8e1393c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.373297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.768s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.373947] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.377074] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.551s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.378632] env[62066]: INFO nova.compute.claims [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.392332] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155830, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069279} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.392583] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.393609] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984ec113-da86-420e-966f-ea196ae5510b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.417412] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.419036] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e1c4a6-dd55-4dc9-af50-50a0082fae43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.441657] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 733.441657] env[62066]: value = "task-1155831" [ 733.441657] env[62066]: _type = "Task" [ 733.441657] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.451273] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155831, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.618300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8de5cea-a145-4957-9ba2-f4e9d0ca909c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.638803] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 733.672697] env[62066]: DEBUG oslo_vmware.api [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155827, 'name': PowerOnVM_Task, 'duration_secs': 0.864629} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.672972] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 733.672972] env[62066]: INFO nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 9.34 seconds to spawn the instance on the hypervisor. [ 733.673288] env[62066]: DEBUG nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 733.674113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826414c1-117d-486b-b540-826b91f20632 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.738550] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155828, 'name': ReconfigVM_Task, 'duration_secs': 0.62888} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.738679] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 265906be-704e-4f0f-ae33-e89135e27268/265906be-704e-4f0f-ae33-e89135e27268.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.739300] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5755d40-f76b-4934-b507-7b06a4c28010 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.746500] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 733.746500] env[62066]: value = "task-1155832" [ 733.746500] env[62066]: _type = "Task" [ 733.746500] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.754702] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155832, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.852939] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c94ca9-b537-4a48-70a0-b6eb8e1393c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009751} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.853297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.853644] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 733.853955] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9dd8dad1-3844-409d-b2c1-c9b587ece91f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.862286] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 733.862286] env[62066]: value = "task-1155833" [ 733.862286] env[62066]: _type = "Task" [ 733.862286] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.871921] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.887965] env[62066]: DEBUG nova.compute.utils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 733.890432] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.893740] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.952215] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155831, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.978513] env[62066]: DEBUG nova.policy [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '852509ce34bd4391ad7ab2a0e5d70a1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70d709f95cfa418392c56eb730c053c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.147455] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.147455] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-feef7bbb-6502-4cfc-b0a3-1ac76314700c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.155856] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 734.155856] env[62066]: value = "task-1155834" [ 734.155856] env[62066]: _type = "Task" [ 734.155856] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.169382] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.197079] env[62066]: INFO nova.compute.manager [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 28.12 seconds to build instance. [ 734.259034] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155832, 'name': Rename_Task, 'duration_secs': 0.157348} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.259442] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 734.259751] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c24773a3-d0bd-45ec-89d6-5ba86ec6a8dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.268855] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 734.268855] env[62066]: value = "task-1155835" [ 734.268855] env[62066]: _type = "Task" [ 734.268855] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.277847] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155835, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.375738] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155833, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.394389] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.456525] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155831, 'name': ReconfigVM_Task, 'duration_secs': 0.527105} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.459710] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.460887] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b17b3cd6-e32b-426f-b387-af7e05fd1d30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.467202] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 734.467202] env[62066]: value = "task-1155836" [ 734.467202] env[62066]: _type = "Task" [ 734.467202] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.476576] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155836, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.627309] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Successfully created port: e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.666302] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155834, 'name': PowerOffVM_Task, 'duration_secs': 0.303556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.669199] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.669527] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 734.701554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d18388b-3d06-4307-ba83-6db8ae177c2b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.964s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.781465] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155835, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.837294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b851fefe-52c3-4c86-ad0b-58980a9b109f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.844886] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45e81b6-c42d-4a44-aece-ee97e16ad101 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.884187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de72b8-ae87-44b9-b4b9-9426dcf3d09f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.892365] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155833, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542426} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.894877] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 734.895761] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4369b5-55d4-4f51-8c55-f346cf99ee1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.899507] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee45d84-2553-4df2-b520-f035ef52232d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.921243] env[62066]: DEBUG nova.compute.provider_tree [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.944230] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 734.944230] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-145996b5-a62d-439f-87bd-2905944cbe1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.963817] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 734.963817] env[62066]: value = "task-1155837" [ 734.963817] env[62066]: _type = "Task" [ 734.963817] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.974181] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155837, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.981730] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155836, 'name': Rename_Task, 'duration_secs': 0.230875} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.982053] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 734.982302] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f74fb119-0714-438b-86e3-1d2e237a8cad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.988138] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 734.988138] env[62066]: value = "task-1155838" [ 734.988138] env[62066]: _type = "Task" [ 734.988138] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.996098] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.176915] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.177411] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.177778] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.178140] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.178506] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.178885] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.179278] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.179471] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.179708] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.183116] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.183116] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.185264] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-648c34aa-e294-40d5-adb9-c8951f7b32e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.202793] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 735.202793] env[62066]: value = "task-1155839" [ 735.202793] env[62066]: _type = "Task" [ 735.202793] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.207344] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.215253] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155839, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.284728] env[62066]: DEBUG oslo_vmware.api [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155835, 'name': PowerOnVM_Task, 'duration_secs': 0.516986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.285050] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.285271] env[62066]: INFO nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Took 14.51 seconds to spawn the instance on the hypervisor. [ 735.285474] env[62066]: DEBUG nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.286644] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d06039-d602-4c5b-9a1b-06b533dbacb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.411132] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.442365] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.442738] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.442983] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.443299] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.443574] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.443769] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.444107] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.444386] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.444635] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.444919] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.445231] env[62066]: DEBUG nova.virt.hardware [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.447857] env[62066]: DEBUG nova.scheduler.client.report [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.454151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e331d72f-1759-4c12-964e-18a14d15fb86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.464455] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f40cfa6-d793-4796-a70b-f1c82fd4d93b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.487526] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.498396] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155838, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.558227] env[62066]: DEBUG nova.compute.manager [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Received event network-changed-81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.558282] env[62066]: DEBUG nova.compute.manager [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Refreshing instance network info cache due to event network-changed-81a78519-507f-4de8-8e08-3a469f901c85. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.558491] env[62066]: DEBUG oslo_concurrency.lockutils [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.558645] env[62066]: DEBUG oslo_concurrency.lockutils [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.558824] env[62066]: DEBUG nova.network.neutron [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Refreshing network info cache for port 81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 735.717497] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155839, 'name': ReconfigVM_Task, 'duration_secs': 0.237515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.717807] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 735.749563] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.808923] env[62066]: INFO nova.compute.manager [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Took 36.12 seconds to build instance. [ 735.960675] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.960675] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.962916] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.498s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.964353] env[62066]: INFO nova.compute.claims [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.980033] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155837, 'name': ReconfigVM_Task, 'duration_secs': 0.770978} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.980033] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Reconfigured VM instance instance-0000002e to attach disk [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.980750] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f97695-37c5-4f3e-bfa0-60b4beb2e425 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.010319] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88ee08dd-c184-46a2-973a-e2d0583de906 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.031258] env[62066]: DEBUG oslo_vmware.api [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155838, 'name': PowerOnVM_Task, 'duration_secs': 0.755678} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.033047] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 736.033047] env[62066]: INFO nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Took 7.04 seconds to spawn the instance on the hypervisor. [ 736.033239] env[62066]: DEBUG nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.033576] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 736.033576] env[62066]: value = "task-1155840" [ 736.033576] env[62066]: _type = "Task" [ 736.033576] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.034301] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccf2ba1-a68f-4d97-a9c1-97327bb372e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.045993] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155840, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.225143] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:53:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f58b5fca-fe02-4466-adf4-baeb8793402a',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-738942274',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.225451] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.225663] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.225868] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.226058] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.226227] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.226447] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.226623] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.226854] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.226972] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.227178] env[62066]: DEBUG nova.virt.hardware [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.232994] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfiguring VM instance instance-00000028 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 736.233743] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584eb305-bf67-494c-a5bb-59e74419ce58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.256132] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 736.256132] env[62066]: value = "task-1155841" [ 736.256132] env[62066]: _type = "Task" [ 736.256132] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.265138] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155841, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.314782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4dffb828-3f98-4d63-b3c1-265b91a063dc tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.231s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.334474] env[62066]: DEBUG nova.network.neutron [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updated VIF entry in instance network info cache for port 81a78519-507f-4de8-8e08-3a469f901c85. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 736.334890] env[62066]: DEBUG nova.network.neutron [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.469215] env[62066]: DEBUG nova.compute.utils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 736.472641] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.472890] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.519137] env[62066]: DEBUG nova.compute.manager [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-vif-plugged-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.519137] env[62066]: DEBUG oslo_concurrency.lockutils [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] Acquiring lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.519137] env[62066]: DEBUG oslo_concurrency.lockutils [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.519137] env[62066]: DEBUG oslo_concurrency.lockutils [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.519454] env[62066]: DEBUG nova.compute.manager [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] No waiting events found dispatching network-vif-plugged-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 736.519547] env[62066]: WARNING nova.compute.manager [req-58365d03-cd58-4873-aa1d-12fa25af763e req-0283bd0c-fab8-4b79-ac2d-fa2523c3d8de service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received unexpected event network-vif-plugged-e8c586f5-3024-40d6-934b-04b551be3c44 for instance with vm_state building and task_state spawning. [ 736.522820] env[62066]: DEBUG nova.policy [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f453531ab9864e8ba056c5cf2dbfb603', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '120bc7d4eb004ee3982fad8283f6f885', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.546690] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155840, 'name': ReconfigVM_Task, 'duration_secs': 0.19534} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.546983] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.551461] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a745bf9-0833-49ba-a1fb-0966ab4528ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.563201] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 736.563201] env[62066]: value = "task-1155842" [ 736.563201] env[62066]: _type = "Task" [ 736.563201] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.567467] env[62066]: INFO nova.compute.manager [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Took 27.99 seconds to build instance. [ 736.577017] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155842, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.594683] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Successfully updated port: e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.771155] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155841, 'name': ReconfigVM_Task, 'duration_secs': 0.208456} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.771487] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfigured VM instance instance-00000028 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 736.772358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a46db59-e88a-42a6-b619-886b4ee17cae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.807118] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.807523] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f210a0-840f-40f2-b320-80ecfaa02b9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.821418] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Successfully created port: 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.823462] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.831330] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 736.831330] env[62066]: value = "task-1155843" [ 736.831330] env[62066]: _type = "Task" [ 736.831330] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.838249] env[62066]: DEBUG oslo_concurrency.lockutils [req-c2a861fd-ec72-4c4e-8700-253fe645b062 req-d39dc9a7-37ed-4caa-b6a1-66bd6509286f service nova] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.842232] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.976204] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.069682] env[62066]: DEBUG oslo_concurrency.lockutils [None req-922ac799-e962-47f3-84b4-103d2c8b6bd6 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.450s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.074525] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155842, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.099417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.099417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.099417] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 737.122729] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 737.123755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c00801-a226-4a05-b77e-d6cf52e4fd8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.131115] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 737.131245] env[62066]: ERROR oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk due to incomplete transfer. [ 737.133642] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ae4a9cc8-3890-4661-b6b0-81c5f488f295 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.141153] env[62066]: DEBUG oslo_vmware.rw_handles [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52fc9fc6-da32-1558-f3f7-1a02d400ad5a/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 737.141362] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Uploaded image 218d1044-e6e2-4a2e-b8ab-4d2ff641ed5b to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 737.143778] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 737.146179] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6a7f7b80-6925-4ccc-a1b5-96a276af2d59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.152536] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 737.152536] env[62066]: value = "task-1155844" [ 737.152536] env[62066]: _type = "Task" [ 737.152536] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.162402] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155844, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.347244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.359195] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155843, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.385445] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331327a7-0d69-4f06-a880-704c87e35bff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.394630] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3293be4a-07ed-4777-bc8d-ca1c2718a133 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.432382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.432710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.432962] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "265906be-704e-4f0f-ae33-e89135e27268-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.433175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.433352] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.435714] env[62066]: INFO nova.compute.manager [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Terminating instance [ 737.437592] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f03d555-bbb6-4913-9b85-d8b993dcf57d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.440631] env[62066]: DEBUG nova.compute.manager [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.440860] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.441571] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81848290-f6d0-41e3-b3ac-44ddd335aefe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.453804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde2cfe3-d3f6-46a3-9138-55697ca9446c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.456825] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.457109] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a22e4dd-5fb2-4f2c-aba3-c507089fa0e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.468926] env[62066]: DEBUG nova.compute.provider_tree [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.471263] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 737.471263] env[62066]: value = "task-1155845" [ 737.471263] env[62066]: _type = "Task" [ 737.471263] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.479493] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.574625] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155842, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.576157] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.658076] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.666195] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155844, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.844162] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155843, 'name': ReconfigVM_Task, 'duration_secs': 0.63008} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.844477] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.844737] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 737.973460] env[62066]: DEBUG nova.scheduler.client.report [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.989970] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.994025] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155845, 'name': PowerOffVM_Task, 'duration_secs': 0.208123} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.994025] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 737.994025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 737.994025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0818ad82-0096-4b5c-be32-e0a637b6640c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.010687] env[62066]: INFO nova.compute.manager [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Rebuilding instance [ 738.018106] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.018384] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.018548] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.018757] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.018923] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.019086] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.019355] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.019487] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.019688] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.019870] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.020057] env[62066]: DEBUG nova.virt.hardware [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.020934] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca870325-9830-4dd0-b394-4f224b77acb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.030257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2d1429-0674-466e-9f6f-6c8b72728e92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.076039] env[62066]: DEBUG oslo_vmware.api [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155842, 'name': PowerOnVM_Task, 'duration_secs': 1.097028} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.076487] env[62066]: DEBUG nova.compute.manager [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.077648] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 738.081505] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd27af45-360e-421f-a6cb-09acf4e93258 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.082188] env[62066]: DEBUG nova.compute.manager [None req-ccee0c21-c8a5-492f-8343-039eb1618192 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.084765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfee8ce-58b9-4a63-b3b1-cb579423a945 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.125868] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.127524] env[62066]: DEBUG nova.network.neutron [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.130602] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 738.130602] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 738.130602] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleting the datastore file [datastore1] 265906be-704e-4f0f-ae33-e89135e27268 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 738.130602] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39f06d9f-2c48-46c3-976f-d942f38c28df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.143028] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 738.143028] env[62066]: value = "task-1155847" [ 738.143028] env[62066]: _type = "Task" [ 738.143028] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.154318] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155847, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.164440] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155844, 'name': Destroy_Task, 'duration_secs': 0.612449} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.164755] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Destroyed the VM [ 738.165126] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 738.165463] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ed85b608-150b-464f-9ffb-196b1c735f51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.175034] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 738.175034] env[62066]: value = "task-1155848" [ 738.175034] env[62066]: _type = "Task" [ 738.175034] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.182677] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155848, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.351653] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df60b85e-285a-45de-91d5-9a539549d518 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.375081] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4d08ea-de20-42b4-8b6a-a8c2068d4802 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.395461] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 738.401014] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Successfully updated port: 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 738.426742] env[62066]: DEBUG nova.compute.manager [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received event network-vif-plugged-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.427071] env[62066]: DEBUG oslo_concurrency.lockutils [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] Acquiring lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.427319] env[62066]: DEBUG oslo_concurrency.lockutils [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] Lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.427900] env[62066]: DEBUG oslo_concurrency.lockutils [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] Lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.428044] env[62066]: DEBUG nova.compute.manager [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] No waiting events found dispatching network-vif-plugged-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 738.428211] env[62066]: WARNING nova.compute.manager [req-0a442191-1e50-4190-b886-7cdd8a77d767 req-5af0f6bf-ce9a-4797-8e67-433f4a6efb93 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received unexpected event network-vif-plugged-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d for instance with vm_state building and task_state spawning. [ 738.485403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.486055] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.489138] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.590s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.490726] env[62066]: INFO nova.compute.claims [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.602700] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 738.605808] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7db74d2f-bd31-4e88-a0d2-e388cbab57ff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.615947] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 738.615947] env[62066]: value = "task-1155849" [ 738.615947] env[62066]: _type = "Task" [ 738.615947] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.627417] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.632759] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.633260] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Instance network_info: |[{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 738.634100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:b9:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9ee6f9-33be-4f58-8248-694024ec31d4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8c586f5-3024-40d6-934b-04b551be3c44', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 738.643345] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Creating folder: Project (70d709f95cfa418392c56eb730c053c3). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.644157] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-049e093e-bbf7-44f7-bbb9-1b4b574dd03a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.653358] env[62066]: DEBUG nova.compute.manager [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.653437] env[62066]: DEBUG nova.compute.manager [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing instance network info cache due to event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.653669] env[62066]: DEBUG oslo_concurrency.lockutils [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.653815] env[62066]: DEBUG oslo_concurrency.lockutils [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.654063] env[62066]: DEBUG nova.network.neutron [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.661451] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Created folder: Project (70d709f95cfa418392c56eb730c053c3) in parent group-v251573. [ 738.661652] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Creating folder: Instances. Parent ref: group-v251623. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.664743] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cd7f258-c476-4bec-8343-cc86b071fc17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.666756] env[62066]: DEBUG oslo_vmware.api [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1155847, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163058} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.666987] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.667186] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 738.667372] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.667539] env[62066]: INFO nova.compute.manager [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Took 1.23 seconds to destroy the instance on the hypervisor. [ 738.667773] env[62066]: DEBUG oslo.service.loopingcall [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.668320] env[62066]: DEBUG nova.compute.manager [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.668414] env[62066]: DEBUG nova.network.neutron [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.679538] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Created folder: Instances in parent group-v251623. [ 738.679773] env[62066]: DEBUG oslo.service.loopingcall [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.680196] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 738.680773] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b18bfab3-677c-473f-83fb-7dfa4b852a63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.698858] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155848, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.704189] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 738.704189] env[62066]: value = "task-1155852" [ 738.704189] env[62066]: _type = "Task" [ 738.704189] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.712275] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155852, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.908376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.908727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.908727] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.973556] env[62066]: DEBUG nova.network.neutron [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Port 4627c549-cbf1-40f9-94bb-0f08255025ae binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 738.998794] env[62066]: DEBUG nova.compute.utils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 738.999865] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.000094] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.071207] env[62066]: DEBUG nova.policy [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a698e0440d54b5786cf04eab68fa375', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34b4eb3e6a6a402e90edcfa4958bf6eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.128250] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155849, 'name': PowerOffVM_Task, 'duration_secs': 0.235474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.128588] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 739.128773] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.129592] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef3aded-bba6-48f2-b40d-cec2b7ed46ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.136524] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 739.136787] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acd79b25-9e8c-4aec-afe1-5d7baf5c85cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.191889] env[62066]: DEBUG oslo_vmware.api [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155848, 'name': RemoveSnapshot_Task, 'duration_secs': 0.94734} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.191889] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 739.191889] env[62066]: INFO nova.compute.manager [None req-de24c177-24f5-4d11-b50e-6fefb0ba33c3 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 16.65 seconds to snapshot the instance on the hypervisor. [ 739.203804] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 739.204123] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 739.204285] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.205395] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f53375f5-1046-4f80-8fa4-2df9cbd7dd9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.951961] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Successfully created port: 4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.962882] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.976176] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 739.976176] env[62066]: value = "task-1155854" [ 739.976176] env[62066]: _type = "Task" [ 739.976176] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.977154] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155852, 'name': CreateVM_Task, 'duration_secs': 0.338714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.977154] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 739.980801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.981015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.981368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 739.981669] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-375a1e4c-c8b5-4395-aab5-89eb3df6e149 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.988440] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 739.988440] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5201a1c6-d6d1-c1e1-117c-cfa0dab1cd93" [ 739.988440] env[62066]: _type = "Task" [ 739.988440] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.992287] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155854, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.004214] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5201a1c6-d6d1-c1e1-117c-cfa0dab1cd93, 'name': SearchDatastore_Task, 'duration_secs': 0.011415} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.009489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.009768] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.010025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.010188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.010420] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.010940] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39cea2da-cb45-4a6b-962d-3b0d3802e58c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.020525] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.020738] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.022131] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e53e31b-7087-4d80-bb81-6b6a1b09a53a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.033258] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 740.033258] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52911c94-9501-ffcc-9a5e-b49da481f9e0" [ 740.033258] env[62066]: _type = "Task" [ 740.033258] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.034917] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.048460] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52911c94-9501-ffcc-9a5e-b49da481f9e0, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.049679] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b731d8a0-768b-4b9b-9af4-2ecf2988effa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.055736] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 740.055736] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52667d2b-08b9-8d52-1e03-5b6db0623f83" [ 740.055736] env[62066]: _type = "Task" [ 740.055736] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.071323] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52667d2b-08b9-8d52-1e03-5b6db0623f83, 'name': SearchDatastore_Task, 'duration_secs': 0.00935} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.074233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.074465] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 82e2a71a-d27f-4db9-8f84-16762d3d3bf6/82e2a71a-d27f-4db9-8f84-16762d3d3bf6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 740.077023] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-888ed8e7-f1f8-42e4-beba-e79ebdfcca0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.085212] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 740.085212] env[62066]: value = "task-1155855" [ 740.085212] env[62066]: _type = "Task" [ 740.085212] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.094571] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.267203] env[62066]: DEBUG nova.network.neutron [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updated VIF entry in instance network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 740.267574] env[62066]: DEBUG nova.network.neutron [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.385945] env[62066]: DEBUG nova.network.neutron [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.455791] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9712c6-d99e-4307-8a53-c9e881df37d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.477981] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817558d9-ce97-4e00-90aa-e55c0847e704 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.488597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.489053] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.489053] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.500651] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155854, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162308} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.528652] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.528652] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 740.528652] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.531110] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c22129-0a35-406d-9e5f-ab4f5156dd04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.539476] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6bef20-5d32-472e-9c94-19f0667ea55b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.559344] env[62066]: DEBUG nova.compute.provider_tree [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.595475] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48147} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.599040] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 82e2a71a-d27f-4db9-8f84-16762d3d3bf6/82e2a71a-d27f-4db9-8f84-16762d3d3bf6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 740.599040] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.599040] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-151168f6-3bce-44c1-9d65-0d7f4e86df94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.604501] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 740.604501] env[62066]: value = "task-1155856" [ 740.604501] env[62066]: _type = "Task" [ 740.604501] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.613173] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.764593] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.764792] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing instance network info cache due to event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.765260] env[62066]: DEBUG oslo_concurrency.lockutils [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] Acquiring lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.772566] env[62066]: DEBUG oslo_concurrency.lockutils [req-fa10452e-7b25-45ad-b7a1-a624f0de0ff3 req-05829633-8167-4e55-81a4-41a4d9f6d376 service nova] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.889680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.890741] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Instance network_info: |[{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 740.890741] env[62066]: DEBUG oslo_concurrency.lockutils [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] Acquired lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.890741] env[62066]: DEBUG nova.network.neutron [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.892442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:80:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.900811] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Creating folder: Project (120bc7d4eb004ee3982fad8283f6f885). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.900921] env[62066]: INFO nova.compute.manager [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Rescuing [ 740.901125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.901281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.901450] env[62066]: DEBUG nova.network.neutron [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.903390] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45861e4b-8c62-4343-9a58-1ada7e1d5888 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.915155] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Created folder: Project (120bc7d4eb004ee3982fad8283f6f885) in parent group-v251573. [ 740.915353] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Creating folder: Instances. Parent ref: group-v251626. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.915593] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49b84b26-020e-4c12-b5de-ecf15ed71093 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.927845] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Created folder: Instances in parent group-v251626. [ 740.928114] env[62066]: DEBUG oslo.service.loopingcall [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.928313] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.928559] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ec425ff-426e-46b7-b53f-6fc025c4ef69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.947902] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.947902] env[62066]: value = "task-1155859" [ 740.947902] env[62066]: _type = "Task" [ 740.947902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.955540] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155859, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.985809] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.012667] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.012840] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.012935] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.013101] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.013253] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.013401] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.013608] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.013766] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.013932] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.014110] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.014291] env[62066]: DEBUG nova.virt.hardware [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.015386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd79b73-658c-4da1-b94e-925defbf2dce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.023233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cb126d-66ed-46c7-8d1a-497ecb497854 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.058078] env[62066]: DEBUG nova.scheduler.client.report [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.061967] env[62066]: DEBUG nova.network.neutron [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.115421] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066444} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.115958] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 741.116888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a8e839-f4be-4bff-8d4f-7c068d5c6ce9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.140407] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 82e2a71a-d27f-4db9-8f84-16762d3d3bf6/82e2a71a-d27f-4db9-8f84-16762d3d3bf6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 741.141211] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd1ad1c6-1ed9-451d-9521-33fe86a4188e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.163116] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 741.163116] env[62066]: value = "task-1155860" [ 741.163116] env[62066]: _type = "Task" [ 741.163116] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.171309] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155860, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.308669] env[62066]: DEBUG nova.compute.manager [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 741.309718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1757733-2a4f-4aa9-a84f-1e7a9fc978e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.458422] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155859, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.565713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.566357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.567852] env[62066]: DEBUG nova.network.neutron [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.572554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.081s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.572554] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.573711] env[62066]: INFO nova.compute.manager [-] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Took 2.91 seconds to deallocate network for instance. [ 741.579417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.199s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.584295] env[62066]: INFO nova.compute.claims [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.607041] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.607672] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.607964] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.608462] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.608850] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.609269] env[62066]: DEBUG nova.virt.hardware [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.611403] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39bdba3-265e-4582-9561-90412185cc38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.629743] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabf9f1a-e752-4e69-8891-fc4b89b94811 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.650734] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:16:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd18922bc-f1a1-4553-8a99-0ec96b254d2d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.662321] env[62066]: DEBUG oslo.service.loopingcall [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.663842] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 741.664141] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-121b7fe0-5e5e-4c54-9b12-6d8a993e7a41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.690349] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.694103] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 741.694103] env[62066]: value = "task-1155861" [ 741.694103] env[62066]: _type = "Task" [ 741.694103] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.708619] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155861, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.735057] env[62066]: DEBUG nova.network.neutron [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updated VIF entry in instance network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.735594] env[62066]: DEBUG nova.network.neutron [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.819095] env[62066]: DEBUG nova.network.neutron [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [{"id": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "address": "fa:16:3e:70:9f:6f", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4fe3c0-de", "ovs_interfaceid": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.821053] env[62066]: INFO nova.compute.manager [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] instance snapshotting [ 741.823941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c93a695-7b22-45e9-a378-979dddcdeae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.847151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d19a7d-af2b-4f42-9239-ebbf203f2d03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.960867] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155859, 'name': CreateVM_Task, 'duration_secs': 0.807419} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.961165] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 741.962961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.962961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.962961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.962961] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4b4f0d6-3235-42aa-88bf-b9b9ffdef674 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.967509] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 741.967509] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52751297-b52a-7bb5-83f6-cb8c307edbea" [ 741.967509] env[62066]: _type = "Task" [ 741.967509] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.975700] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52751297-b52a-7bb5-83f6-cb8c307edbea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.075035] env[62066]: DEBUG nova.compute.utils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 742.076568] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.076745] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.096953] env[62066]: DEBUG nova.compute.manager [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Received event network-vif-plugged-4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.100561] env[62066]: DEBUG oslo_concurrency.lockutils [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] Acquiring lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.100561] env[62066]: DEBUG oslo_concurrency.lockutils [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.100561] env[62066]: DEBUG oslo_concurrency.lockutils [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.100561] env[62066]: DEBUG nova.compute.manager [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] No waiting events found dispatching network-vif-plugged-4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 742.100561] env[62066]: WARNING nova.compute.manager [req-3dbabeca-c472-48e4-99d9-207e3afa595d req-aa54a668-4d44-4163-8bd6-eb0d6124e45e service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Received unexpected event network-vif-plugged-4e4bf2f1-d884-4bd4-91b5-8d1574133055 for instance with vm_state building and task_state spawning. [ 742.101375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.174846] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155860, 'name': ReconfigVM_Task, 'duration_secs': 0.746445} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.176722] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 82e2a71a-d27f-4db9-8f84-16762d3d3bf6/82e2a71a-d27f-4db9-8f84-16762d3d3bf6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.179016] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70e339a8-95b7-4684-b3d4-1c3ff29abe9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.183944] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 742.183944] env[62066]: value = "task-1155862" [ 742.183944] env[62066]: _type = "Task" [ 742.183944] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.196243] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155862, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.198069] env[62066]: DEBUG nova.policy [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e276807bc06444eae42968bd7528926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a8e27dbc37043b3b36f8492c47d2286', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.206402] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Successfully updated port: 4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 742.215023] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155861, 'name': CreateVM_Task, 'duration_secs': 0.504806} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.215023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 742.215023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.215023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.215023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 742.215023] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c679420d-fd75-49fb-a1c7-2b0d702e1b14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.219640] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 742.219640] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bc7b0c-89b9-fab5-1d7e-646c0ac95070" [ 742.219640] env[62066]: _type = "Task" [ 742.219640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.229691] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bc7b0c-89b9-fab5-1d7e-646c0ac95070, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.238280] env[62066]: DEBUG oslo_concurrency.lockutils [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] Releasing lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.238553] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-deleted-77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.238737] env[62066]: INFO nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Neutron deleted interface 77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9; detaching it from the instance and deleting it from the info cache [ 742.239049] env[62066]: DEBUG nova.network.neutron [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [{"id": "ab2ecd7c-9249-477c-aeef-56105dc12410", "address": "fa:16:3e:86:4d:f1", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2ecd7c-92", "ovs_interfaceid": "ab2ecd7c-9249-477c-aeef-56105dc12410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "address": "fa:16:3e:b6:32:d9", "network": {"id": "f05688d3-a5c8-4a5a-85cc-ce8d7dcde861", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1381993909", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.168", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5729cb45-fe", "ovs_interfaceid": "5729cb45-fe23-4d92-b6a2-8b5f9164d4e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.324427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.361386] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 742.361875] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9e2c623e-4081-45a7-89f1-74c08402b493 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.364578] env[62066]: DEBUG nova.network.neutron [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.371277] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 742.371277] env[62066]: value = "task-1155863" [ 742.371277] env[62066]: _type = "Task" [ 742.371277] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.379832] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155863, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.479114] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52751297-b52a-7bb5-83f6-cb8c307edbea, 'name': SearchDatastore_Task, 'duration_secs': 0.015478} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.479564] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.480212] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.480503] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.480677] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.480866] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.481152] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0012efe9-1c1f-4039-b3fc-2bc0cc5f4168 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.488628] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.488808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 742.489580] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1d74b80-c767-40c1-88fa-c6aa61fa20db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.494279] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 742.494279] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5223b136-0686-c964-a861-a894a5772446" [ 742.494279] env[62066]: _type = "Task" [ 742.494279] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.503338] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5223b136-0686-c964-a861-a894a5772446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.579762] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.699596] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155862, 'name': Rename_Task, 'duration_secs': 0.145614} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.699950] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 742.700192] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80ad9d36-0e40-40b9-b9f0-ebbcaad34072 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.707998] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 742.707998] env[62066]: value = "task-1155864" [ 742.707998] env[62066]: _type = "Task" [ 742.707998] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.715579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.715579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquired lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.715579] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.717252] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.734363] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bc7b0c-89b9-fab5-1d7e-646c0ac95070, 'name': SearchDatastore_Task, 'duration_secs': 0.009569} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.734757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.735052] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.735898] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.735898] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.736050] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.736553] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-515674ae-ccd2-4bd4-80a0-0170954bbbb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.743329] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-161b3694-cc9e-47a1-b3e3-c0e30d81ca00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.749116] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.749563] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 742.752070] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbe0f682-3353-4e73-8075-bde3b6d6d94a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.763272] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644ed048-d655-4c9c-b33c-4c70e741a398 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.785246] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 742.785246] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523a4b6f-2955-aba5-105c-aa8352839874" [ 742.785246] env[62066]: _type = "Task" [ 742.785246] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.805461] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Detach interface failed, port_id=77b2e9c0-f2ef-4e58-bf7a-959c8aee4aa9, reason: Instance 265906be-704e-4f0f-ae33-e89135e27268 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 742.805874] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-deleted-5729cb45-fe23-4d92-b6a2-8b5f9164d4e7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.806193] env[62066]: INFO nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Neutron deleted interface 5729cb45-fe23-4d92-b6a2-8b5f9164d4e7; detaching it from the instance and deleting it from the info cache [ 742.806601] env[62066]: DEBUG nova.network.neutron [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [{"id": "ab2ecd7c-9249-477c-aeef-56105dc12410", "address": "fa:16:3e:86:4d:f1", "network": {"id": "01cc2bf6-ef62-43db-bf0e-7e13e8386adb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-950739290", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.36", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab2ecd7c-92", "ovs_interfaceid": "ab2ecd7c-9249-477c-aeef-56105dc12410", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.819821] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523a4b6f-2955-aba5-105c-aa8352839874, 'name': SearchDatastore_Task, 'duration_secs': 0.011542} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.820238] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-351785c7-b371-4604-acee-73837695df04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.829227] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 742.829227] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd908-b03e-9eed-033a-08e978ca7968" [ 742.829227] env[62066]: _type = "Task" [ 742.829227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.837685] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd908-b03e-9eed-033a-08e978ca7968, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.867396] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.871221] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.872603] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b71c9e1-bbad-41a4-b34b-4bcc0a02e47d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.881714] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155863, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.882994] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 742.882994] env[62066]: value = "task-1155865" [ 742.882994] env[62066]: _type = "Task" [ 742.882994] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.895682] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.917369] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Successfully created port: 7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.943292] env[62066]: DEBUG nova.compute.manager [req-aefa858c-5bb2-4bcf-9fa8-50593e0e77da req-ffbfff8a-d287-488c-aba8-af6e82f44fbb service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Received event network-vif-deleted-ab2ecd7c-9249-477c-aeef-56105dc12410 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.943689] env[62066]: INFO nova.compute.manager [req-aefa858c-5bb2-4bcf-9fa8-50593e0e77da req-ffbfff8a-d287-488c-aba8-af6e82f44fbb service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Neutron deleted interface ab2ecd7c-9249-477c-aeef-56105dc12410; detaching it from the instance and deleting it from the info cache [ 742.943982] env[62066]: DEBUG nova.network.neutron [req-aefa858c-5bb2-4bcf-9fa8-50593e0e77da req-ffbfff8a-d287-488c-aba8-af6e82f44fbb service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.009018] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5223b136-0686-c964-a861-a894a5772446, 'name': SearchDatastore_Task, 'duration_secs': 0.007896} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.009293] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5c6e9c4-2546-4678-b104-7ffc6bd0091f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.016047] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 743.016047] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522a4153-aad9-c9b1-d430-0fcb6bdec07c" [ 743.016047] env[62066]: _type = "Task" [ 743.016047] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.026712] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522a4153-aad9-c9b1-d430-0fcb6bdec07c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.103261] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e55d1b-0a9f-4efb-94b9-7e5ad7f2880d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.111700] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d54959d-2ec1-426e-9ec4-1d6d068c9a25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.146604] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654efa94-5f6c-4dcd-97e0-0b42dbfe6809 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.155754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29239b98-03bc-454c-ade1-dec5aebeb556 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.172155] env[62066]: DEBUG nova.compute.provider_tree [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.219948] env[62066]: DEBUG oslo_vmware.api [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1155864, 'name': PowerOnVM_Task, 'duration_secs': 0.456527} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.220251] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 743.220491] env[62066]: INFO nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Took 7.81 seconds to spawn the instance on the hypervisor. [ 743.220676] env[62066]: DEBUG nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.221755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d140619d-86a3-48eb-8c10-d2ed97ecb31c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.277561] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.315532] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-539b40be-221d-4989-b1d5-ec9dc4f8b57b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.325208] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c06bbe-d55b-47fd-83f9-8477395d8eb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.347651] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd908-b03e-9eed-033a-08e978ca7968, 'name': SearchDatastore_Task, 'duration_secs': 0.009524} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.348038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.348247] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.348506] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1308c4f3-8ab8-49ad-9f45-a06f43a87421 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.359411] env[62066]: DEBUG nova.compute.manager [req-2137a638-5921-4c43-9895-9a4f5fc3053c req-275dce5c-b4f3-43e6-a69b-58f34fe3b3fe service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Detach interface failed, port_id=5729cb45-fe23-4d92-b6a2-8b5f9164d4e7, reason: Instance 265906be-704e-4f0f-ae33-e89135e27268 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 743.363665] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 743.363665] env[62066]: value = "task-1155866" [ 743.363665] env[62066]: _type = "Task" [ 743.363665] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.371729] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.388165] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155863, 'name': CreateSnapshot_Task, 'duration_secs': 0.794393} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.391357] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 743.392173] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc62fffa-9bc8-496d-99ed-ba0a685e9286 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.396207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05a205d-8e6b-4a38-bb1c-6d2471812d8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.401597] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155865, 'name': PowerOffVM_Task, 'duration_secs': 0.16011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.402286] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.403128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d8a712-eed3-4ff5-99aa-edaa2ebeffd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.428650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ce2ba4-4857-4d38-b342-3a31e8ea5b23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.448167] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-055f329e-438b-4533-9787-af973f601b5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.450955] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958f19cb-ee8c-440d-8cee-a4cf145c1d7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.455808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 743.470262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4920a97d-a40d-4033-b19d-078238c62791 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.499573] env[62066]: DEBUG nova.compute.manager [req-aefa858c-5bb2-4bcf-9fa8-50593e0e77da req-ffbfff8a-d287-488c-aba8-af6e82f44fbb service nova] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Detach interface failed, port_id=ab2ecd7c-9249-477c-aeef-56105dc12410, reason: Instance 265906be-704e-4f0f-ae33-e89135e27268 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 743.501756] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 743.502152] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0384bf7e-28f7-48e3-8dd9-12c13feed2d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.508338] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 743.508338] env[62066]: value = "task-1155867" [ 743.508338] env[62066]: _type = "Task" [ 743.508338] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.512360] env[62066]: DEBUG nova.network.neutron [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Updating instance_info_cache with network_info: [{"id": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "address": "fa:16:3e:e7:17:d6", "network": {"id": "3673c3e2-3ed0-4fed-9d48-d998bb6bcf81", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-210535853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34b4eb3e6a6a402e90edcfa4958bf6eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e4bf2f1-d8", "ovs_interfaceid": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.522911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 743.523186] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.523401] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.529628] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522a4153-aad9-c9b1-d430-0fcb6bdec07c, 'name': SearchDatastore_Task, 'duration_secs': 0.010268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.530019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.530172] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 641bf537-76d2-4c41-a6dc-b5b042327477/641bf537-76d2-4c41-a6dc-b5b042327477.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.530500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.530746] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.531060] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d7427a3-c7ba-497a-b618-a8ad11b91554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.533051] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aab8f376-b4f8-471f-85ac-bb12062b4ede {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.540259] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 743.540259] env[62066]: value = "task-1155868" [ 743.540259] env[62066]: _type = "Task" [ 743.540259] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.544802] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.544952] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 743.546110] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-137a85b8-8004-48d1-b795-2e241b767b94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.552694] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.555070] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 743.555070] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e5f47d-db97-b7ad-ffe9-2fe5a54b6f4d" [ 743.555070] env[62066]: _type = "Task" [ 743.555070] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.563874] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e5f47d-db97-b7ad-ffe9-2fe5a54b6f4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.591177] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.616163] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.616882] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.616882] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.616882] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.617148] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.617300] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.617438] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.617628] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.617778] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.617946] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.618207] env[62066]: DEBUG nova.virt.hardware [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.619198] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc0eaa9b-6d8e-40a1-ada0-15324e63a110 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.627765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04636c81-ab90-460c-a354-8824fbc827c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.677031] env[62066]: DEBUG nova.scheduler.client.report [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.740310] env[62066]: INFO nova.compute.manager [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Took 29.30 seconds to build instance. [ 743.875216] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469099} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.875560] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 743.875865] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.876467] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0ed5522-026a-4385-843f-96e4d2b35824 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.886654] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 743.886654] env[62066]: value = "task-1155869" [ 743.886654] env[62066]: _type = "Task" [ 743.886654] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.898758] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.941192] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 743.941693] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-71b03cb1-1fdf-417f-a95e-7c5f4d44636a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.952271] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 743.952271] env[62066]: value = "task-1155870" [ 743.952271] env[62066]: _type = "Task" [ 743.952271] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.964976] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.964976] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.964976] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae62c500-a3e6-488f-8821-c64bbecfdc80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.971679] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 743.971679] env[62066]: value = "task-1155871" [ 743.971679] env[62066]: _type = "Task" [ 743.971679] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.982086] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.016728] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Releasing lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.016728] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Instance network_info: |[{"id": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "address": "fa:16:3e:e7:17:d6", "network": {"id": "3673c3e2-3ed0-4fed-9d48-d998bb6bcf81", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-210535853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34b4eb3e6a6a402e90edcfa4958bf6eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e4bf2f1-d8", "ovs_interfaceid": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 744.017370] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:17:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e4bf2f1-d884-4bd4-91b5-8d1574133055', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.030937] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Creating folder: Project (34b4eb3e6a6a402e90edcfa4958bf6eb). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.031823] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3067d693-b857-4038-9950-709d088063a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.048705] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Created folder: Project (34b4eb3e6a6a402e90edcfa4958bf6eb) in parent group-v251573. [ 744.049128] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Creating folder: Instances. Parent ref: group-v251632. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.050231] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ef8f716-80f6-487a-b296-c3a2f160c297 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.058834] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155868, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.068864] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Created folder: Instances in parent group-v251632. [ 744.069315] env[62066]: DEBUG oslo.service.loopingcall [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.076062] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.077060] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e5f47d-db97-b7ad-ffe9-2fe5a54b6f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.018098} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.077060] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4e5fc36-23fc-4b37-bdba-28728e95faa1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.095674] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-591d675c-34bc-45ad-ac55-176d23bb8f41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.103696] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.103696] env[62066]: value = "task-1155874" [ 744.103696] env[62066]: _type = "Task" [ 744.103696] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.105183] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 744.105183] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52043c6f-e2fd-4401-4e61-3e328ed7bee9" [ 744.105183] env[62066]: _type = "Task" [ 744.105183] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.118434] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155874, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.122069] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52043c6f-e2fd-4401-4e61-3e328ed7bee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.133208] env[62066]: DEBUG nova.compute.manager [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Received event network-changed-4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.133531] env[62066]: DEBUG nova.compute.manager [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Refreshing instance network info cache due to event network-changed-4e4bf2f1-d884-4bd4-91b5-8d1574133055. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.133860] env[62066]: DEBUG oslo_concurrency.lockutils [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] Acquiring lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.134060] env[62066]: DEBUG oslo_concurrency.lockutils [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] Acquired lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.134280] env[62066]: DEBUG nova.network.neutron [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Refreshing network info cache for port 4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.185780] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.186590] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.191455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.250s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.192056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.192285] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 744.192658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.144s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.194369] env[62066]: INFO nova.compute.claims [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.198875] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2a8393-ce0a-4df4-8624-472a2368fb2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.213063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fcf65d-611b-4dbd-bd73-03bb2723b294 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.232138] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759e4fd8-b975-48bf-a32c-c95c8678ed9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.237879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324f8f8c-1ae5-43e2-aceb-ad57853b7788 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.244634] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0182899-d053-4805-8b4a-ac672511150b tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.746s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.272814] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181520MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 744.273083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.401847] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099847} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.402396] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.403444] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767b6447-14cd-414a-966e-29d5a2a39aa8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.428574] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.428940] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dc0d721-eaee-4fa9-8a44-94a0e6e5f76c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.448783] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 744.448783] env[62066]: value = "task-1155875" [ 744.448783] env[62066]: _type = "Task" [ 744.448783] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.463541] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155875, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.468379] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.481086] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155871, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.550669] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704476} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.550982] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 641bf537-76d2-4c41-a6dc-b5b042327477/641bf537-76d2-4c41-a6dc-b5b042327477.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.551214] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.551466] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63c828ab-d1ad-49de-8785-02cac55f0c9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.558772] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 744.558772] env[62066]: value = "task-1155876" [ 744.558772] env[62066]: _type = "Task" [ 744.558772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.566993] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.617845] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155874, 'name': CreateVM_Task, 'duration_secs': 0.373985} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.621174] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 744.621504] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52043c6f-e2fd-4401-4e61-3e328ed7bee9, 'name': SearchDatastore_Task, 'duration_secs': 0.062619} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.622192] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.622942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.622942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.622942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.623198] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 744.623492] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26ac9b56-c2f0-4af6-8c0c-6cda2a2ef214 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.624980] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e31dc6d-086f-4586-8069-25866b62272d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.629987] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 744.629987] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52348a57-c85a-601f-c27b-294943439021" [ 744.629987] env[62066]: _type = "Task" [ 744.629987] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.633979] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 744.633979] env[62066]: value = "task-1155877" [ 744.633979] env[62066]: _type = "Task" [ 744.633979] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.644092] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52348a57-c85a-601f-c27b-294943439021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.645455] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155877, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.699763] env[62066]: DEBUG nova.compute.utils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 744.701364] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.701583] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 744.772522] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.816810] env[62066]: DEBUG nova.policy [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53366fea3a6434fadadede81df16089', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33b756ddd30f4cb0b917fad171213266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.864303] env[62066]: DEBUG nova.network.neutron [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Updated VIF entry in instance network info cache for port 4e4bf2f1-d884-4bd4-91b5-8d1574133055. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.864766] env[62066]: DEBUG nova.network.neutron [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Updating instance_info_cache with network_info: [{"id": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "address": "fa:16:3e:e7:17:d6", "network": {"id": "3673c3e2-3ed0-4fed-9d48-d998bb6bcf81", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-210535853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34b4eb3e6a6a402e90edcfa4958bf6eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e4bf2f1-d8", "ovs_interfaceid": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.969082] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.976801] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155875, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.987458] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155871, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.075482] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061681} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.076110] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.077151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2773ce-b1bb-42ce-a645-9b6af17e4b1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.112937] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 641bf537-76d2-4c41-a6dc-b5b042327477/641bf537-76d2-4c41-a6dc-b5b042327477.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.113583] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9cd751f-5ee6-42e1-a581-5108b33c414f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.141594] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 745.141594] env[62066]: value = "task-1155878" [ 745.141594] env[62066]: _type = "Task" [ 745.141594] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.147195] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52348a57-c85a-601f-c27b-294943439021, 'name': SearchDatastore_Task, 'duration_secs': 0.012518} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.152529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.152612] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.153504] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.153504] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.153504] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.154975] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155877, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.154975] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Successfully updated port: 7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 745.158717] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dee84b76-661f-4d26-a450-cd801fb22865 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.166108] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.171099] env[62066]: DEBUG nova.compute.manager [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.171307] env[62066]: DEBUG nova.compute.manager [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing instance network info cache due to event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 745.171542] env[62066]: DEBUG oslo_concurrency.lockutils [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.171678] env[62066]: DEBUG oslo_concurrency.lockutils [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.171885] env[62066]: DEBUG nova.network.neutron [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 745.175064] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.175521] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.176556] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6ad78d4-ccb7-4769-8bac-a6e370c5de8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.183827] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 745.183827] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5278c9b5-1261-af05-fdcd-2642b904119e" [ 745.183827] env[62066]: _type = "Task" [ 745.183827] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.194203] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5278c9b5-1261-af05-fdcd-2642b904119e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.205442] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.293435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.367614] env[62066]: DEBUG oslo_concurrency.lockutils [req-a5072297-822e-4a1e-8f75-4c9abb54eb0f req-3d22e14f-3591-4e91-b9e7-135ee4eb671a service nova] Releasing lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.464218] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155875, 'name': ReconfigVM_Task, 'duration_secs': 0.616677} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.465311] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4/6a498481-b2ca-4813-87b7-2f09dfa107f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.466157] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2d2af75-06ff-4fda-8413-6f439fa83cfe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.470859] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.482490] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 745.482490] env[62066]: value = "task-1155879" [ 745.482490] env[62066]: _type = "Task" [ 745.482490] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.487888] env[62066]: DEBUG oslo_vmware.api [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155871, 'name': PowerOnVM_Task, 'duration_secs': 1.027924} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.488511] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.488774] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed0e1ac-e54a-4099-8ec5-305f7850a011 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance 'c0ac4362-766f-48ba-aeb2-7fd976c1f47f' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 745.499088] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155879, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.517773] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Successfully created port: 58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.580172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89fc2ec-8194-470f-a926-b2ed3f8179f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.589777] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a65d57-fbfc-4e97-a46f-70e2b1d7649d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.629765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67213702-c6ec-403d-a0dd-7500ff9c532d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.647961] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7362ce3b-e295-4d78-9b9a-663eb376e7b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.652232] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155877, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525849} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.655342] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 745.656419] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9476755c-4952-4a16-bb57-52794f2fa609 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.666983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.667196] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.667349] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.668693] env[62066]: DEBUG nova.compute.provider_tree [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.674013] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.701111] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.706064] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a9f984f-49c9-477f-a76c-81dcfa7378b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.729845] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5278c9b5-1261-af05-fdcd-2642b904119e, 'name': SearchDatastore_Task, 'duration_secs': 0.016574} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.731916] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 745.731916] env[62066]: value = "task-1155880" [ 745.731916] env[62066]: _type = "Task" [ 745.731916] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.732285] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eaa2534-9e24-44fa-b32c-f119f31f2a18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.746612] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155880, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.747057] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 745.747057] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52845175-df53-3663-c4b5-006cb439d3af" [ 745.747057] env[62066]: _type = "Task" [ 745.747057] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.758167] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52845175-df53-3663-c4b5-006cb439d3af, 'name': SearchDatastore_Task, 'duration_secs': 0.009888} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.759079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.759386] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2/16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 745.759879] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35282465-f382-4fc3-9561-41e93c9a1c6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.767111] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 745.767111] env[62066]: value = "task-1155881" [ 745.767111] env[62066]: _type = "Task" [ 745.767111] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.776021] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.967842] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task} progress is 95%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.993772] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155879, 'name': Rename_Task, 'duration_secs': 0.281027} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.993772] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 745.993772] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dea8b4a-978e-4ac0-8bc4-c0c536c1d2d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.006158] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 746.006158] env[62066]: value = "task-1155882" [ 746.006158] env[62066]: _type = "Task" [ 746.006158] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.026916] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.166795] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155878, 'name': ReconfigVM_Task, 'duration_secs': 0.878587} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.167329] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 641bf537-76d2-4c41-a6dc-b5b042327477/641bf537-76d2-4c41-a6dc-b5b042327477.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.176587] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d4d6897-6d72-4767-94a3-7f7485d60343 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.176655] env[62066]: DEBUG nova.scheduler.client.report [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.186384] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 746.186384] env[62066]: value = "task-1155883" [ 746.186384] env[62066]: _type = "Task" [ 746.186384] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.197543] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155883, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.220974] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.228022] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.246593] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.267020] env[62066]: DEBUG nova.virt.hardware [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.267020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482f0f04-be52-4857-b784-e0510a351635 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.282250] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457997} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.284909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12661cbf-b33e-4e24-96c7-f4999894eb75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.288046] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2/16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 746.288277] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.288568] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2de2caf7-163b-453d-b1d4-856eaac0cc18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.304985] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 746.304985] env[62066]: value = "task-1155884" [ 746.304985] env[62066]: _type = "Task" [ 746.304985] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.313146] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155884, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.374046] env[62066]: DEBUG nova.network.neutron [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updated VIF entry in instance network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 746.374642] env[62066]: DEBUG nova.network.neutron [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.382831] env[62066]: DEBUG nova.network.neutron [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updating instance_info_cache with network_info: [{"id": "7d4d5170-fd58-4398-b403-887f40e3c29c", "address": "fa:16:3e:39:73:e9", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4d5170-fd", "ovs_interfaceid": "7d4d5170-fd58-4398-b403-887f40e3c29c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.391549] env[62066]: DEBUG nova.compute.manager [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Received event network-vif-plugged-7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.394105] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Acquiring lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.394105] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.394105] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.394105] env[62066]: DEBUG nova.compute.manager [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] No waiting events found dispatching network-vif-plugged-7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 746.394105] env[62066]: WARNING nova.compute.manager [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Received unexpected event network-vif-plugged-7d4d5170-fd58-4398-b403-887f40e3c29c for instance with vm_state building and task_state spawning. [ 746.394105] env[62066]: DEBUG nova.compute.manager [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Received event network-changed-7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.394105] env[62066]: DEBUG nova.compute.manager [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Refreshing instance network info cache due to event network-changed-7d4d5170-fd58-4398-b403-887f40e3c29c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.394105] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Acquiring lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.470267] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155870, 'name': CloneVM_Task, 'duration_secs': 2.207473} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.470267] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Created linked-clone VM from snapshot [ 746.470267] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac0e210-9c7e-4467-b2be-9ab46d216c5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.477730] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Uploading image 28a813de-c3d3-4db4-ae6d-742016b9dd6e {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 746.499115] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 746.499115] env[62066]: value = "vm-251631" [ 746.499115] env[62066]: _type = "VirtualMachine" [ 746.499115] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 746.500039] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a70f242d-f96e-47c5-b9d7-6cee0df2a644 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.506029] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease: (returnval){ [ 746.506029] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c201ea-1066-9f63-8c97-7931d4f177c1" [ 746.506029] env[62066]: _type = "HttpNfcLease" [ 746.506029] env[62066]: } obtained for exporting VM: (result){ [ 746.506029] env[62066]: value = "vm-251631" [ 746.506029] env[62066]: _type = "VirtualMachine" [ 746.506029] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 746.506367] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the lease: (returnval){ [ 746.506367] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c201ea-1066-9f63-8c97-7931d4f177c1" [ 746.506367] env[62066]: _type = "HttpNfcLease" [ 746.506367] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 746.519172] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 746.519172] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c201ea-1066-9f63-8c97-7931d4f177c1" [ 746.519172] env[62066]: _type = "HttpNfcLease" [ 746.519172] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 746.522379] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 746.522379] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c201ea-1066-9f63-8c97-7931d4f177c1" [ 746.522379] env[62066]: _type = "HttpNfcLease" [ 746.522379] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 746.522578] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.523258] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35310b27-d76c-422e-85a9-44969514a34a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.530804] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 746.530950] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 746.635381] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-83c15f4f-6655-4457-92b3-a780ef72bd46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.682880] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.683532] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.686668] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.301s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.688321] env[62066]: INFO nova.compute.claims [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.702290] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155883, 'name': Rename_Task, 'duration_secs': 0.173644} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.702290] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 746.702290] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b140eedc-7b17-42df-be7b-7ddecec2c2d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.708193] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 746.708193] env[62066]: value = "task-1155886" [ 746.708193] env[62066]: _type = "Task" [ 746.708193] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.717124] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.746719] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.817048] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155884, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.241023} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.817647] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.818580] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c429b9-1a8d-44f2-88c5-e34c132a8b1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.841669] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2/16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.842163] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43ee630d-aedb-4a87-a225-1ff609511e6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.866033] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 746.866033] env[62066]: value = "task-1155887" [ 746.866033] env[62066]: _type = "Task" [ 746.866033] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.872279] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155887, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.878341] env[62066]: DEBUG oslo_concurrency.lockutils [req-2cf83b58-a32a-4693-bc5e-305372662711 req-ce9ed3e6-3d48-4d93-84b1-4e7def7661fa service nova] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.885257] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.885769] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Instance network_info: |[{"id": "7d4d5170-fd58-4398-b403-887f40e3c29c", "address": "fa:16:3e:39:73:e9", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4d5170-fd", "ovs_interfaceid": "7d4d5170-fd58-4398-b403-887f40e3c29c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 746.886260] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Acquired lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.886626] env[62066]: DEBUG nova.network.neutron [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Refreshing network info cache for port 7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.888281] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:73:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d4d5170-fd58-4398-b403-887f40e3c29c', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.897404] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating folder: Project (8a8e27dbc37043b3b36f8492c47d2286). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.899593] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca94311a-3f11-4fb2-a895-9db37a022c0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.912049] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created folder: Project (8a8e27dbc37043b3b36f8492c47d2286) in parent group-v251573. [ 746.912049] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating folder: Instances. Parent ref: group-v251635. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.912049] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7aea09fc-04a5-4f20-929e-ec59db7e95db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.925362] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created folder: Instances in parent group-v251635. [ 746.926455] env[62066]: DEBUG oslo.service.loopingcall [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.926455] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.926455] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a46725d-dfdb-466c-8bc4-b4d6d5c793c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.946246] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.946246] env[62066]: value = "task-1155890" [ 746.946246] env[62066]: _type = "Task" [ 746.946246] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.955016] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155890, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.016479] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.195780] env[62066]: DEBUG nova.compute.utils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 747.198471] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.198471] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 747.220351] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155886, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.248545] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.261649] env[62066]: DEBUG nova.policy [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5caad5da140d457d88c40c304dd4f605', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab8a3b607934456b5d408442f03524a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.379521] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155887, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.459687] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155890, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.518544] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.619089] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Successfully created port: 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.705024] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.720515] env[62066]: DEBUG oslo_vmware.api [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155886, 'name': PowerOnVM_Task, 'duration_secs': 0.906717} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.721150] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 747.721646] env[62066]: INFO nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Took 9.73 seconds to spawn the instance on the hypervisor. [ 747.722188] env[62066]: DEBUG nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.726413] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe004c0-c937-4fb5-a093-a8fd1dcd9824 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.748678] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155880, 'name': ReconfigVM_Task, 'duration_secs': 1.559538} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.750731] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfigured VM instance instance-0000002d to attach disk [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.751654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29c0c1a-7544-48fa-8dc9-ab6912064444 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.786483] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5f4d697-fe7a-4729-9c1f-a43e4b295b53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.808166] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 747.808166] env[62066]: value = "task-1155891" [ 747.808166] env[62066]: _type = "Task" [ 747.808166] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.828009] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.877968] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155887, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.958844] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155890, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.017753] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.214667] env[62066]: DEBUG nova.network.neutron [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updated VIF entry in instance network info cache for port 7d4d5170-fd58-4398-b403-887f40e3c29c. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 748.214978] env[62066]: DEBUG nova.network.neutron [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updating instance_info_cache with network_info: [{"id": "7d4d5170-fd58-4398-b403-887f40e3c29c", "address": "fa:16:3e:39:73:e9", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4d5170-fd", "ovs_interfaceid": "7d4d5170-fd58-4398-b403-887f40e3c29c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.257175] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Successfully updated port: 58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 748.260259] env[62066]: INFO nova.compute.manager [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Took 29.45 seconds to build instance. [ 748.280615] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a4a2c3-5f3b-4deb-aefa-444f4fd2bbdf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.290960] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dcc994b-3175-42e6-a59e-67ce18d41108 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.327100] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56913fa0-3282-404d-9055-0df3c9c94f23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.338033] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155891, 'name': ReconfigVM_Task, 'duration_secs': 0.231561} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.338184] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 748.338560] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-740e7dc6-7ee7-4802-aa76-b861625e4bb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.341336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02d9880-81e8-406b-ac64-e827aaee8d59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.358059] env[62066]: DEBUG nova.compute.provider_tree [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.360978] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 748.360978] env[62066]: value = "task-1155892" [ 748.360978] env[62066]: _type = "Task" [ 748.360978] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.374532] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155892, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.379762] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155887, 'name': ReconfigVM_Task, 'duration_secs': 1.332907} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.380221] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2/16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.380939] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3d74f0c-a07c-4c16-ad08-5357dbdcd20a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.387687] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 748.387687] env[62066]: value = "task-1155893" [ 748.387687] env[62066]: _type = "Task" [ 748.387687] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.397310] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155893, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.458720] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155890, 'name': CreateVM_Task, 'duration_secs': 1.233152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.459237] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 748.459773] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.459938] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.460304] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.460672] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daefb34e-6b9d-49a7-a370-b642505ad75e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.467443] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 748.467443] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bec100-e17e-7d9a-8d62-eb6a3e0b797d" [ 748.467443] env[62066]: _type = "Task" [ 748.467443] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.478803] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bec100-e17e-7d9a-8d62-eb6a3e0b797d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.519982] env[62066]: DEBUG oslo_vmware.api [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155882, 'name': PowerOnVM_Task, 'duration_secs': 2.455909} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.520686] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.520829] env[62066]: DEBUG nova.compute.manager [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.521708] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22112db-8b82-475c-b4de-c7b86c183fee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.715585] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.718324] env[62066]: DEBUG oslo_concurrency.lockutils [req-b631ff73-c655-404a-88d1-6545c4437e88 req-8f2b4488-8425-4401-af15-5c75df54f32d service nova] Releasing lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.735517] env[62066]: DEBUG nova.compute.manager [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Received event network-vif-plugged-58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.735879] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Acquiring lock "248f5281-eb46-4be1-8642-28813c4b2622-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.736217] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Lock "248f5281-eb46-4be1-8642-28813c4b2622-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.736450] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Lock "248f5281-eb46-4be1-8642-28813c4b2622-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.736684] env[62066]: DEBUG nova.compute.manager [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] No waiting events found dispatching network-vif-plugged-58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 748.736958] env[62066]: WARNING nova.compute.manager [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Received unexpected event network-vif-plugged-58c12e67-83e3-4f98-bbc7-b90eef01f919 for instance with vm_state building and task_state spawning. [ 748.737282] env[62066]: DEBUG nova.compute.manager [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Received event network-changed-58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.737611] env[62066]: DEBUG nova.compute.manager [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Refreshing instance network info cache due to event network-changed-58c12e67-83e3-4f98-bbc7-b90eef01f919. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.737893] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Acquiring lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.738255] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Acquired lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.738555] env[62066]: DEBUG nova.network.neutron [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Refreshing network info cache for port 58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 748.750594] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.750922] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.751201] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.751449] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.751681] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.751879] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.752167] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.752414] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.752772] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.753011] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.753204] env[62066]: DEBUG nova.virt.hardware [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.754361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c3598f-bab3-4b37-ba80-3525890c3b6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.759584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.762549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-edb39deb-3247-4a7a-9f01-e2f93289ad6a tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.217s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.766277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4967dd-7389-4a23-8072-4a5ccc51f8c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.863459] env[62066]: DEBUG nova.scheduler.client.report [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.876320] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155892, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.898514] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155893, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.985591] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bec100-e17e-7d9a-8d62-eb6a3e0b797d, 'name': SearchDatastore_Task, 'duration_secs': 0.011284} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.985967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.990129] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.990482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.990732] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.990996] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.991572] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a38926cc-6554-4408-a635-41c08e17d44f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.002177] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.003187] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 749.003187] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a5997e8-32f4-4101-9834-0fb0176389b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.008577] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 749.008577] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b4964d-1d62-7f9d-d52e-91a2bd53c654" [ 749.008577] env[62066]: _type = "Task" [ 749.008577] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.017620] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b4964d-1d62-7f9d-d52e-91a2bd53c654, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.041354] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.252067] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Successfully updated port: 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.269791] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.354323] env[62066]: DEBUG nova.network.neutron [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.372860] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.373372] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.376088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.627s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.377602] env[62066]: INFO nova.compute.claims [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.386028] env[62066]: DEBUG oslo_vmware.api [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155892, 'name': PowerOnVM_Task, 'duration_secs': 0.610428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.386283] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.388866] env[62066]: DEBUG nova.compute.manager [None req-d0d661bd-04d8-4e6f-a45c-36bd0ab03679 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.390495] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e690155-5351-4bd1-9cd3-92e84455a137 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.405729] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155893, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.411736] env[62066]: DEBUG nova.network.neutron [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Port 4627c549-cbf1-40f9-94bb-0f08255025ae binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 749.412039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.412165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.412333] env[62066]: DEBUG nova.network.neutron [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.450401] env[62066]: DEBUG nova.compute.manager [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.450789] env[62066]: DEBUG oslo_concurrency.lockutils [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.451086] env[62066]: DEBUG oslo_concurrency.lockutils [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.451361] env[62066]: DEBUG oslo_concurrency.lockutils [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.451485] env[62066]: DEBUG nova.compute.manager [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] No waiting events found dispatching network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 749.452335] env[62066]: WARNING nova.compute.manager [req-28ecda95-555b-43a4-8c7b-6c3ea2438529 req-d2a0b1a5-d55a-4fde-a233-28a529057de6 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received unexpected event network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d for instance with vm_state building and task_state spawning. [ 749.520602] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b4964d-1d62-7f9d-d52e-91a2bd53c654, 'name': SearchDatastore_Task, 'duration_secs': 0.010242} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.521361] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c863c752-9c3b-4435-95bc-af7cbe947011 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.526530] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 749.526530] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5203ee68-cc15-d32c-042b-21be65c7d58a" [ 749.526530] env[62066]: _type = "Task" [ 749.526530] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.536259] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5203ee68-cc15-d32c-042b-21be65c7d58a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.575475] env[62066]: DEBUG nova.network.neutron [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.756221] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.756221] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.756221] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.795923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.882503] env[62066]: DEBUG nova.compute.utils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 749.886865] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.887105] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.900375] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155893, 'name': Rename_Task, 'duration_secs': 1.208672} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.900699] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.900997] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80b377ce-2403-4680-ab14-3b31874dfac8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.913772] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 749.913772] env[62066]: value = "task-1155894" [ 749.913772] env[62066]: _type = "Task" [ 749.913772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.930951] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155894, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.986242] env[62066]: DEBUG nova.policy [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91585bdab75c42178c835de912fd3fef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08737e043ede43b58298d1c1a834fa84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.038883] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5203ee68-cc15-d32c-042b-21be65c7d58a, 'name': SearchDatastore_Task, 'duration_secs': 0.012447} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.039417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.043325] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/de9eb96d-e71b-4878-83b8-f75966dc3f48.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 750.043325] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46bea714-15ee-4712-b42f-e3c5fa8dcd6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.049162] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 750.049162] env[62066]: value = "task-1155895" [ 750.049162] env[62066]: _type = "Task" [ 750.049162] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.059162] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.078489] env[62066]: DEBUG oslo_concurrency.lockutils [req-061ecec3-00dc-42ec-af59-612a773fdb45 req-8552dab1-5f27-4a07-9273-5379da719249 service nova] Releasing lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.078952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.079110] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.292583] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.387583] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.433704] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155894, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.483423] env[62066]: DEBUG nova.network.neutron [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.524621] env[62066]: DEBUG nova.network.neutron [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.571425] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155895, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.650659] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.867017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7fcb6a-58a1-4734-9f63-1722f68eeae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.874613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa16b3c-96ea-498c-8981-984673cb2b77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.914972] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Successfully created port: 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.919433] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2f6ba1-119e-4d7d-8f91-f64c4fdac93c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.934109] env[62066]: DEBUG oslo_vmware.api [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155894, 'name': PowerOnVM_Task, 'duration_secs': 0.703193} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.935351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11cd625-46fc-4034-9961-aa5b6b23fbef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.940120] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.941922] env[62066]: INFO nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Took 9.95 seconds to spawn the instance on the hypervisor. [ 750.942192] env[62066]: DEBUG nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.943107] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d91b278-05ef-490a-a956-21ba54d051eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.958162] env[62066]: DEBUG nova.compute.provider_tree [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.986535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.986913] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance network_info: |[{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 750.987383] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:29:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7b4bfde-f109-4f64-adab-e7f06b80685d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4991ab48-c8b7-437b-81d7-db93c7d3da4d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 750.995980] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating folder: Project (7ab8a3b607934456b5d408442f03524a). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 750.996664] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a821c31f-6cbd-4863-8694-ad3d5f1b159c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.007956] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Created folder: Project (7ab8a3b607934456b5d408442f03524a) in parent group-v251573. [ 751.008210] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating folder: Instances. Parent ref: group-v251638. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.008473] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1306a370-e818-4cce-bb89-3129af357fd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.017595] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Created folder: Instances in parent group-v251638. [ 751.017870] env[62066]: DEBUG oslo.service.loopingcall [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.018110] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.018335] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6eb59ae-de73-4bf6-b002-138366cd2f65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.036490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.042368] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.042368] env[62066]: value = "task-1155898" [ 751.042368] env[62066]: _type = "Task" [ 751.042368] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.050785] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155898, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.061127] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608026} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.061917] env[62066]: DEBUG nova.network.neutron [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Updating instance_info_cache with network_info: [{"id": "58c12e67-83e3-4f98-bbc7-b90eef01f919", "address": "fa:16:3e:9e:18:74", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58c12e67-83", "ovs_interfaceid": "58c12e67-83e3-4f98-bbc7-b90eef01f919", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.063196] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/de9eb96d-e71b-4878-83b8-f75966dc3f48.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.063519] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.063984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a306f9c-2b4a-4bb9-aea8-316368e07c64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.074473] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 751.074473] env[62066]: value = "task-1155899" [ 751.074473] env[62066]: _type = "Task" [ 751.074473] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.085713] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.184969] env[62066]: DEBUG nova.compute.manager [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.185231] env[62066]: DEBUG nova.compute.manager [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing instance network info cache due to event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.185509] env[62066]: DEBUG oslo_concurrency.lockutils [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] Acquiring lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.185714] env[62066]: DEBUG oslo_concurrency.lockutils [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] Acquired lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.185896] env[62066]: DEBUG nova.network.neutron [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.190851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "641bf537-76d2-4c41-a6dc-b5b042327477" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.190990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.191266] env[62066]: INFO nova.compute.manager [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Rebooting instance [ 751.245603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.245945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.246291] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.246533] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.246745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.249606] env[62066]: INFO nova.compute.manager [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Terminating instance [ 751.251603] env[62066]: DEBUG nova.compute.manager [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.251876] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.252769] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7ea48f-2107-4ea2-bd64-2f14c88d763a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.260901] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 751.261196] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40a38e24-9563-431a-a285-11e2d0bd0ec9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.268859] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 751.268859] env[62066]: value = "task-1155900" [ 751.268859] env[62066]: _type = "Task" [ 751.268859] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.278878] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.425635] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.452494] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.452983] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.452983] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.453207] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.453369] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.453553] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.453812] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.454462] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.454462] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.455719] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.455719] env[62066]: DEBUG nova.virt.hardware [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.456021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7a5644-03f9-4902-bf02-936a09ff3c7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.468038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d72c06-b078-4a7c-80c5-125c66ea641b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.471437] env[62066]: DEBUG nova.scheduler.client.report [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.492211] env[62066]: INFO nova.compute.manager [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Took 30.05 seconds to build instance. [ 751.540130] env[62066]: DEBUG nova.compute.manager [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62066) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 751.540449] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.553639] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155898, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.567618] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-248f5281-eb46-4be1-8642-28813c4b2622" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.567966] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance network_info: |[{"id": "58c12e67-83e3-4f98-bbc7-b90eef01f919", "address": "fa:16:3e:9e:18:74", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58c12e67-83", "ovs_interfaceid": "58c12e67-83e3-4f98-bbc7-b90eef01f919", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 751.568408] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:18:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58c12e67-83e3-4f98-bbc7-b90eef01f919', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.582589] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating folder: Project (33b756ddd30f4cb0b917fad171213266). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.583359] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-465e856a-0a6b-436b-8e41-1570a250b048 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.594818] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150526} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.595100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.595889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ea547e-28b3-4201-ba8a-882c64be5b5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.599364] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created folder: Project (33b756ddd30f4cb0b917fad171213266) in parent group-v251573. [ 751.599573] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating folder: Instances. Parent ref: group-v251641. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.600139] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb9e58bd-e59b-431e-ad02-648456742e10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.620189] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/de9eb96d-e71b-4878-83b8-f75966dc3f48.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.620835] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4540a744-64a0-4c64-b183-352db0b218e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.636908] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created folder: Instances in parent group-v251641. [ 751.637200] env[62066]: DEBUG oslo.service.loopingcall [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.637802] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.638053] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2862aa80-1273-4f8c-a769-fe65f101bb16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.654153] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 751.654153] env[62066]: value = "task-1155903" [ 751.654153] env[62066]: _type = "Task" [ 751.654153] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.659167] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.659167] env[62066]: value = "task-1155904" [ 751.659167] env[62066]: _type = "Task" [ 751.659167] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.666436] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.671470] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155904, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.717772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.780024] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155900, 'name': PowerOffVM_Task, 'duration_secs': 0.277042} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.780437] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.780586] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.780807] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1b9a22c-6980-4217-aa67-789904ea891a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.848734] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 751.849019] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 751.849286] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore1] 6a498481-b2ca-4813-87b7-2f09dfa107f4 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.849640] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84acbcf8-a010-40b4-88e8-7cb75d9acf3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.857411] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 751.857411] env[62066]: value = "task-1155906" [ 751.857411] env[62066]: _type = "Task" [ 751.857411] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.866594] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.979913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.980566] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.983621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.638s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.985320] env[62066]: INFO nova.compute.claims [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.999066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5543db67-f2ca-497c-84dd-8ded24a8b7e8 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.222s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.057326] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155898, 'name': CreateVM_Task, 'duration_secs': 0.514519} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.057703] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.058708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.058913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.059302] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.059699] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9360b0-0b20-43ea-ab90-070671e5891a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.065645] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 752.065645] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e53c58-1f81-7aea-31cf-ba6ba5e5fcea" [ 752.065645] env[62066]: _type = "Task" [ 752.065645] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.077529] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e53c58-1f81-7aea-31cf-ba6ba5e5fcea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.167706] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155903, 'name': ReconfigVM_Task, 'duration_secs': 0.411268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.168973] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Reconfigured VM instance instance-00000035 to attach disk [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/de9eb96d-e71b-4878-83b8-f75966dc3f48.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.169744] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a596e6d9-af4f-4c20-a075-81db86ea41d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.176050] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155904, 'name': CreateVM_Task, 'duration_secs': 0.419824} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.176792] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.177479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.181984] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 752.181984] env[62066]: value = "task-1155907" [ 752.181984] env[62066]: _type = "Task" [ 752.181984] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.190540] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155907, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.194891] env[62066]: DEBUG nova.network.neutron [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updated VIF entry in instance network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.195355] env[62066]: DEBUG nova.network.neutron [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.370621] env[62066]: DEBUG oslo_vmware.api [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1155906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199029} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.370621] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.370755] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 752.370885] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.371081] env[62066]: INFO nova.compute.manager [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 752.371345] env[62066]: DEBUG oslo.service.loopingcall [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.371540] env[62066]: DEBUG nova.compute.manager [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.371634] env[62066]: DEBUG nova.network.neutron [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.401576] env[62066]: DEBUG nova.compute.manager [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.401869] env[62066]: DEBUG nova.compute.manager [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing instance network info cache due to event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 752.402136] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.402407] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.402494] env[62066]: DEBUG nova.network.neutron [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.490662] env[62066]: DEBUG nova.compute.utils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 752.496256] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.496256] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 752.500323] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.540221] env[62066]: DEBUG nova.policy [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23cd072660ee4fac8996a2835904531d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db9c97ec5490499bb671cbbae656b3fd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.576498] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e53c58-1f81-7aea-31cf-ba6ba5e5fcea, 'name': SearchDatastore_Task, 'duration_secs': 0.011727} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.576754] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.576998] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.577265] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.577436] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.577656] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.578074] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.578303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.578509] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33137959-a115-43b0-8299-8a2997bf51a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.580683] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84764c1c-9d14-487a-974d-c992d7dff7e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.586872] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 752.586872] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52134ae8-95ae-2ac3-06df-0ed3b4ead1c2" [ 752.586872] env[62066]: _type = "Task" [ 752.586872] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.591655] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.591655] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.593128] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13f3af76-0fb2-4652-9105-e05a59422199 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.598536] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52134ae8-95ae-2ac3-06df-0ed3b4ead1c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.602308] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 752.602308] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52173eb0-f673-9c93-ddde-594cb64a20e5" [ 752.602308] env[62066]: _type = "Task" [ 752.602308] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.612163] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52173eb0-f673-9c93-ddde-594cb64a20e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.693581] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155907, 'name': Rename_Task, 'duration_secs': 0.202389} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.693922] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.694802] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecdf41c1-d592-44fe-9bd3-2ec128c276ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.697938] env[62066]: DEBUG oslo_concurrency.lockutils [req-7888198b-7a3c-4501-beac-d016fb31071b req-45b45e3f-44d2-47a8-9404-f8f33323e9a7 service nova] Releasing lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.699372] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquired lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.699372] env[62066]: DEBUG nova.network.neutron [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.701458] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 752.701458] env[62066]: value = "task-1155908" [ 752.701458] env[62066]: _type = "Task" [ 752.701458] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.709912] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155908, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.816435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.816828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.889028] env[62066]: INFO nova.compute.manager [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Unrescuing [ 752.889028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.889028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.889028] env[62066]: DEBUG nova.network.neutron [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.903203] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Successfully created port: 2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.999406] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.027262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.105134] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52134ae8-95ae-2ac3-06df-0ed3b4ead1c2, 'name': SearchDatastore_Task, 'duration_secs': 0.016758} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.110798] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.111478] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.111836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.125891] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52173eb0-f673-9c93-ddde-594cb64a20e5, 'name': SearchDatastore_Task, 'duration_secs': 0.018733} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.125891] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-480953e5-f91c-4021-8525-8898e2147a9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.136940] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 753.136940] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52401c1f-2a23-cfed-d4d4-5f3fb707b304" [ 753.136940] env[62066]: _type = "Task" [ 753.136940] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.144392] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52401c1f-2a23-cfed-d4d4-5f3fb707b304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.215728] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155908, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.314376] env[62066]: DEBUG nova.network.neutron [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.459450] env[62066]: DEBUG nova.network.neutron [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.533593] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e83945e-5f49-4c04-9bfc-7a0ce48768b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.542665] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9c7801-bd7e-4960-a263-899848a47fda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.581596] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cf73b1-14d0-4f8d-9641-62f7e01ef93d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.592790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a29fbe9-c976-488f-9dc2-a9cf3acc7b61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.614418] env[62066]: DEBUG nova.compute.provider_tree [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.619264] env[62066]: DEBUG nova.compute.manager [req-b402c7be-8fe1-4364-82f3-dc2fb5cd449a req-034de95a-92da-4c85-be47-75137b952be8 service nova] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Received event network-vif-deleted-d18922bc-f1a1-4553-8a99-0ec96b254d2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.646795] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52401c1f-2a23-cfed-d4d4-5f3fb707b304, 'name': SearchDatastore_Task, 'duration_secs': 0.011781} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.647225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.647616] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.648063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.648384] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.648709] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08ad5fe3-371d-4aba-ab12-e5d9ed86be39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.652151] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92ca8483-3b8c-46de-925a-6a19537f0d61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.661653] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 753.661653] env[62066]: value = "task-1155909" [ 753.661653] env[62066]: _type = "Task" [ 753.661653] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.663157] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.663259] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.667153] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f26fbbd-49b7-4c5d-929e-f8875c742179 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.675394] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.677020] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 753.677020] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52897d41-05c0-69a8-af73-ec7fb0f1dbc3" [ 753.677020] env[62066]: _type = "Task" [ 753.677020] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.686037] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52897d41-05c0-69a8-af73-ec7fb0f1dbc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.697351] env[62066]: DEBUG nova.network.neutron [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updated VIF entry in instance network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 753.697783] env[62066]: DEBUG nova.network.neutron [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.713714] env[62066]: DEBUG oslo_vmware.api [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155908, 'name': PowerOnVM_Task, 'duration_secs': 0.768093} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.713958] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.714183] env[62066]: INFO nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Took 10.12 seconds to spawn the instance on the hypervisor. [ 753.714363] env[62066]: DEBUG nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.715182] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e779e06e-4e44-4c17-9749-24c0511158fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.816270] env[62066]: INFO nova.compute.manager [-] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Took 1.44 seconds to deallocate network for instance. [ 753.962369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Releasing lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.968007] env[62066]: DEBUG nova.compute.manager [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.969087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2422843d-1c42-4b24-95c0-b03ab52d62d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.015204] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.046896] env[62066]: DEBUG nova.network.neutron [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [{"id": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "address": "fa:16:3e:70:9f:6f", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4fe3c0-de", "ovs_interfaceid": "3e4fe3c0-de42-48eb-97b5-6f00fafd3a95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.059399] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.060061] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.060061] env[62066]: DEBUG nova.virt.hardware [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.061236] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184c9e69-77b0-405a-8dcd-2615f3735248 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.068217] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Successfully updated port: 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 754.076854] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d5611a-6a0c-47dc-839b-0889da76ab62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.105369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.105369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.105369] env[62066]: INFO nova.compute.manager [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Rebooting instance [ 754.118280] env[62066]: DEBUG nova.scheduler.client.report [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.178450] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155909, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.187640] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52897d41-05c0-69a8-af73-ec7fb0f1dbc3, 'name': SearchDatastore_Task, 'duration_secs': 0.018115} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.188893] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eff4ae7c-dd79-4729-b2ba-1a5a7a36952e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.195179] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 754.195179] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a31e1c-4078-cd7c-8d0b-66026e201881" [ 754.195179] env[62066]: _type = "Task" [ 754.195179] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.200449] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fd00807-bd56-4b3f-aa34-d7b3a194b759 req-a436c4aa-227b-4767-9e26-e7db217f157b service nova] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.204457] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a31e1c-4078-cd7c-8d0b-66026e201881, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.240074] env[62066]: INFO nova.compute.manager [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Took 32.36 seconds to build instance. [ 754.322725] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.501498] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Successfully updated port: 2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 754.552649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-d50f0505-66f4-412f-9744-25e0ea96277a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.553382] env[62066]: DEBUG nova.objects.instance [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'flavor' on Instance uuid d50f0505-66f4-412f-9744-25e0ea96277a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 754.571796] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.571973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.572171] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 754.625076] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.625753] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.630375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.630375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquired lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.630375] env[62066]: DEBUG nova.network.neutron [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 754.630810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.632278] env[62066]: INFO nova.compute.claims [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.674530] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.794497} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.674843] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.675231] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.675515] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-510c7d70-b163-4163-b980-9bd79225a920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.682247] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 754.682247] env[62066]: value = "task-1155910" [ 754.682247] env[62066]: _type = "Task" [ 754.682247] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.691077] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.706153] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a31e1c-4078-cd7c-8d0b-66026e201881, 'name': SearchDatastore_Task, 'duration_secs': 0.058128} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.706475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.706764] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 248f5281-eb46-4be1-8642-28813c4b2622/248f5281-eb46-4be1-8642-28813c4b2622.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.707074] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-039a2268-4043-4e7f-8675-96b7c8f29f47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.714760] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 754.714760] env[62066]: value = "task-1155911" [ 754.714760] env[62066]: _type = "Task" [ 754.714760] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.723382] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.746432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9177a4f1-7ada-448e-8c0a-c28d53d255b1 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.121s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.796621] env[62066]: DEBUG nova.compute.manager [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Received event network-vif-plugged-2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.796850] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] Acquiring lock "1c694323-616c-43b1-bd49-bba707e0788f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.797235] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] Lock "1c694323-616c-43b1-bd49-bba707e0788f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.797599] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] Lock "1c694323-616c-43b1-bd49-bba707e0788f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.797803] env[62066]: DEBUG nova.compute.manager [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] No waiting events found dispatching network-vif-plugged-2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 754.797978] env[62066]: WARNING nova.compute.manager [req-9fcb645b-18d6-43c0-90a2-284f071e68df req-8166d6e9-a92f-4a04-911e-024018070bcf service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Received unexpected event network-vif-plugged-2998edab-7365-4088-a93e-99cbbd0456fe for instance with vm_state building and task_state spawning. [ 754.987300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b9cda0-c61f-4a04-92d4-6f82a9767ffd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.996683] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Doing hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 754.996977] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-019fb21c-14f2-4ae4-9b5e-b72b145b7332 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.002896] env[62066]: DEBUG oslo_vmware.api [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 755.002896] env[62066]: value = "task-1155912" [ 755.002896] env[62066]: _type = "Task" [ 755.002896] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.011770] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.011947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquired lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.012113] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.013381] env[62066]: DEBUG oslo_vmware.api [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155912, 'name': ResetVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.065499] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f0fca9-a65f-4555-9f66-ec6ac39467a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.107056] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.107615] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11bc81e9-d64f-459a-b158-fd03d173a015 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.115580] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 755.115580] env[62066]: value = "task-1155913" [ 755.115580] env[62066]: _type = "Task" [ 755.115580] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.125097] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.140217] env[62066]: DEBUG nova.compute.utils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 755.145143] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.145143] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.148324] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.196238] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074015} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.196238] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.197091] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19b7925-89cc-4272-971c-bebef65e8f2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.228035] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.231303] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63bf1b5d-36d3-4ed0-9baf-c858283ad17a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.249774] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.259468] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155911, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.266839] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 755.266839] env[62066]: value = "task-1155914" [ 755.266839] env[62066]: _type = "Task" [ 755.266839] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.281777] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155914, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.346907] env[62066]: DEBUG nova.policy [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5ca638747b2492181fe6b0f68e683fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb0ca5356fdc4cbca29f97315a25483b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.381100] env[62066]: DEBUG nova.network.neutron [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.517507] env[62066]: DEBUG oslo_vmware.api [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155912, 'name': ResetVM_Task, 'duration_secs': 0.109136} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.518055] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Did hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 755.518279] env[62066]: DEBUG nova.compute.manager [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.519872] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9bfd49-b269-43f0-aaa7-cab10cb54b2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.556336] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.574760] env[62066]: INFO nova.compute.manager [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Rescuing [ 755.574760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.574760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.574760] env[62066]: DEBUG nova.network.neutron [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.626950] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155913, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.645995] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.730800] env[62066]: DEBUG nova.network.neutron [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Updating instance_info_cache with network_info: [{"id": "2998edab-7365-4088-a93e-99cbbd0456fe", "address": "fa:16:3e:21:2e:c5", "network": {"id": "e3c5de25-e78b-49d2-b3c8-3f321eae6068", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1465671414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db9c97ec5490499bb671cbbae656b3fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2998edab-73", "ovs_interfaceid": "2998edab-7365-4088-a93e-99cbbd0456fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.738364] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.802859} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.738854] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 248f5281-eb46-4be1-8642-28813c4b2622/248f5281-eb46-4be1-8642-28813c4b2622.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.742958] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.743322] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90bfb247-02b5-4f17-b3a0-0879c2aa519b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.756821] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 755.756821] env[62066]: value = "task-1155915" [ 755.756821] env[62066]: _type = "Task" [ 755.756821] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.776878] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155915, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.777947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.783613] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155914, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.803662] env[62066]: DEBUG nova.network.neutron [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Updating instance_info_cache with network_info: [{"id": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "address": "fa:16:3e:e7:17:d6", "network": {"id": "3673c3e2-3ed0-4fed-9d48-d998bb6bcf81", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-210535853-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34b4eb3e6a6a402e90edcfa4958bf6eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ed262a3-f84f-4e1c-bbaf-c6a10e0c243e", "external-id": "nsx-vlan-transportzone-146", "segmentation_id": 146, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e4bf2f1-d8", "ovs_interfaceid": "4e4bf2f1-d884-4bd4-91b5-8d1574133055", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.884353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.884839] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Instance network_info: |[{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 755.886119] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:9c:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05f9bc18-80b0-42b9-ba2b-8e552a0a77ab', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.898023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating folder: Project (08737e043ede43b58298d1c1a834fa84). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.898210] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffbf3972-51ab-4283-81bc-800a67b9aaf1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.909991] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created folder: Project (08737e043ede43b58298d1c1a834fa84) in parent group-v251573. [ 755.910248] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating folder: Instances. Parent ref: group-v251644. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.910923] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49e4d543-39cc-41db-bdb7-c51e91809114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.921026] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created folder: Instances in parent group-v251644. [ 755.921026] env[62066]: DEBUG oslo.service.loopingcall [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.921026] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.921026] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5726dde-c11e-479d-8074-b192aff5c554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.945064] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.945064] env[62066]: value = "task-1155918" [ 755.945064] env[62066]: _type = "Task" [ 755.945064] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.954840] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155918, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.041742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b6de4318-2232-4b5e-b62d-bba2cd8942f7 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.851s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.064432] env[62066]: DEBUG nova.compute.manager [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received event network-vif-plugged-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.064822] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Acquiring lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.064822] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.064994] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.065270] env[62066]: DEBUG nova.compute.manager [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] No waiting events found dispatching network-vif-plugged-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 756.065488] env[62066]: WARNING nova.compute.manager [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received unexpected event network-vif-plugged-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab for instance with vm_state building and task_state spawning. [ 756.065682] env[62066]: DEBUG nova.compute.manager [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.065906] env[62066]: DEBUG nova.compute.manager [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing instance network info cache due to event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 756.066182] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Acquiring lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.066372] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Acquired lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.066605] env[62066]: DEBUG nova.network.neutron [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 756.136748] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155913, 'name': PowerOffVM_Task, 'duration_secs': 0.536991} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.137087] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 756.142716] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfiguring VM instance instance-0000002d to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 756.146178] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94a518f0-6d45-45a4-bbae-9332b2380de7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.160988] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Successfully created port: 600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.173150] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 756.173150] env[62066]: value = "task-1155919" [ 756.173150] env[62066]: _type = "Task" [ 756.173150] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.186024] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155919, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.216746] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccea86d-1a9f-45e3-b3c7-45210b640484 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.225877] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebce4ca-8072-4b73-b1be-aede5073c6db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.257630] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Releasing lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.257980] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Instance network_info: |[{"id": "2998edab-7365-4088-a93e-99cbbd0456fe", "address": "fa:16:3e:21:2e:c5", "network": {"id": "e3c5de25-e78b-49d2-b3c8-3f321eae6068", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1465671414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db9c97ec5490499bb671cbbae656b3fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2998edab-73", "ovs_interfaceid": "2998edab-7365-4088-a93e-99cbbd0456fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 756.258989] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:2e:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2998edab-7365-4088-a93e-99cbbd0456fe', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 756.266584] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Creating folder: Project (db9c97ec5490499bb671cbbae656b3fd). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.267420] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a25387a-4d0b-4727-bc1c-d09f4fac1fff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.270393] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6864fc9d-fb31-4908-8960-4be32d0eb0b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.287704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f65398-e550-4f49-b17c-409c0edffe79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.292349] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155915, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.209525} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.296086] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.296574] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Created folder: Project (db9c97ec5490499bb671cbbae656b3fd) in parent group-v251573. [ 756.296791] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Creating folder: Instances. Parent ref: group-v251647. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 756.298489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224b4c4c-7be5-41d4-82cc-89a2a19fc9f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.301923] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2f55248-68bd-4f47-afec-847e34d8885c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.316727] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155914, 'name': ReconfigVM_Task, 'duration_secs': 0.916927} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.317254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Releasing lock "refresh_cache-16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.319303] env[62066]: DEBUG nova.compute.provider_tree [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.320833] env[62066]: DEBUG nova.compute.manager [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.323951] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Reconfigured VM instance instance-00000037 to attach disk [datastore1] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.323951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f67c320-7e67-437f-8ab1-8ccfce67a6f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.335338] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b7c4d85-b039-462e-92a7-e1756ef9e948 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.345797] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 248f5281-eb46-4be1-8642-28813c4b2622/248f5281-eb46-4be1-8642-28813c4b2622.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.348735] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-623623f0-21ad-4369-888b-b2913d6640c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.364462] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Created folder: Instances in parent group-v251647. [ 756.364684] env[62066]: DEBUG oslo.service.loopingcall [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.366896] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 756.370234] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb258e50-0824-434f-ab56-b038009aeb9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.387480] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 756.387480] env[62066]: value = "task-1155923" [ 756.387480] env[62066]: _type = "Task" [ 756.387480] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.387838] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 756.387838] env[62066]: value = "task-1155922" [ 756.387838] env[62066]: _type = "Task" [ 756.387838] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.400532] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 756.400532] env[62066]: value = "task-1155924" [ 756.400532] env[62066]: _type = "Task" [ 756.400532] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.412887] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155922, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.413178] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.419260] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155924, 'name': CreateVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.455805] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155918, 'name': CreateVM_Task, 'duration_secs': 0.469822} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.456224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 756.457217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.457437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.457951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 756.458327] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a190cadc-f9b9-44a4-afa3-8bd18872b91a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.464395] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 756.464395] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526d1dc9-5880-5478-358f-cebc616260a4" [ 756.464395] env[62066]: _type = "Task" [ 756.464395] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.476537] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526d1dc9-5880-5478-358f-cebc616260a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.669260] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.687109] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155919, 'name': ReconfigVM_Task, 'duration_secs': 0.441304} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.687419] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Reconfigured VM instance instance-0000002d to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 756.687706] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.690917] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92b006af-55cf-493b-b6c4-646805e53ab7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.709436] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 756.709436] env[62066]: value = "task-1155925" [ 756.709436] env[62066]: _type = "Task" [ 756.709436] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.712816] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.713182] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.713364] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.713757] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.713757] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.713999] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.714311] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.714506] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.714845] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.715258] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.715368] env[62066]: DEBUG nova.virt.hardware [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.716439] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1a9716-5cdd-4820-9685-7222f831fd05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.733860] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce350f7b-1b9b-4439-b05b-779b7ffe57df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.740402] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155925, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.828682] env[62066]: DEBUG nova.scheduler.client.report [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.914068] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155923, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.920283] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155922, 'name': Rename_Task, 'duration_secs': 0.185429} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.920283] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.920283] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da87570c-a3a4-441c-ac10-7b1d555bfc25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.926856] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155924, 'name': CreateVM_Task, 'duration_secs': 0.441215} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.927519] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 756.928491] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.933493] env[62066]: DEBUG nova.network.neutron [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updated VIF entry in instance network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 756.933493] env[62066]: DEBUG nova.network.neutron [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.935030] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 756.935030] env[62066]: value = "task-1155926" [ 756.935030] env[62066]: _type = "Task" [ 756.935030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.946429] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155926, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.970716] env[62066]: DEBUG nova.network.neutron [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updating instance_info_cache with network_info: [{"id": "7d4d5170-fd58-4398-b403-887f40e3c29c", "address": "fa:16:3e:39:73:e9", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d4d5170-fd", "ovs_interfaceid": "7d4d5170-fd58-4398-b403-887f40e3c29c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.984042] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526d1dc9-5880-5478-358f-cebc616260a4, 'name': SearchDatastore_Task, 'duration_secs': 0.013756} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.985853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.985853] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.985853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.985853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.985853] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.986629] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.986997] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 756.987271] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03e7d733-9f3b-4eb9-89c3-4e6732b6c48c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.989627] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5598c7cc-83d7-412f-ad0e-fbf1b0926963 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.001712] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 757.001712] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5292588a-23c9-74ba-03e5-bc807f156a2c" [ 757.001712] env[62066]: _type = "Task" [ 757.001712] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.001998] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.002376] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.003659] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7de9960-d1e5-4a14-8994-7435583d1de4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.014200] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 757.014200] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aa4645-0a1b-21da-ed25-dbe969036085" [ 757.014200] env[62066]: _type = "Task" [ 757.014200] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.018147] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5292588a-23c9-74ba-03e5-bc807f156a2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.028017] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52aa4645-0a1b-21da-ed25-dbe969036085, 'name': SearchDatastore_Task, 'duration_secs': 0.0101} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.028937] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c58fcf39-fb77-4541-8149-d90d3d0d1338 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.035409] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 757.035409] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524b66d7-a7f8-ea6a-b26f-a8999416ec50" [ 757.035409] env[62066]: _type = "Task" [ 757.035409] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.046247] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524b66d7-a7f8-ea6a-b26f-a8999416ec50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.231589] env[62066]: DEBUG oslo_vmware.api [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1155925, 'name': PowerOnVM_Task, 'duration_secs': 0.492646} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.232079] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.232509] env[62066]: DEBUG nova.compute.manager [None req-7597eb06-27c0-4b1d-a278-5857317eb6ee tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.233827] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f9bb09-4916-4b2b-a5c6-0d2ba351f823 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.335963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.704s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.335963] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.338464] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.238s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.338690] env[62066]: DEBUG nova.objects.instance [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lazy-loading 'resources' on Instance uuid 265906be-704e-4f0f-ae33-e89135e27268 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 757.362433] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 757.364428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e93667-0b07-43e5-b800-de2d00d55207 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.374219] env[62066]: DEBUG nova.compute.manager [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Received event network-changed-2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.374219] env[62066]: DEBUG nova.compute.manager [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Refreshing instance network info cache due to event network-changed-2998edab-7365-4088-a93e-99cbbd0456fe. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.374631] env[62066]: DEBUG oslo_concurrency.lockutils [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] Acquiring lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.374834] env[62066]: DEBUG oslo_concurrency.lockutils [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] Acquired lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.375137] env[62066]: DEBUG nova.network.neutron [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Refreshing network info cache for port 2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.382248] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 757.382541] env[62066]: ERROR oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk due to incomplete transfer. [ 757.382903] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-455c4e2f-13d1-4d19-9d2f-fc34c7c09def {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.401341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212f21dc-7595-4638-94df-0ba175018ab7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.405987] env[62066]: DEBUG oslo_vmware.rw_handles [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52406ee2-a0cb-b83b-5650-9bed1c2a3e6d/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 757.406294] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Uploaded image 28a813de-c3d3-4db4-ae6d-742016b9dd6e to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 757.408984] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 757.409336] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d43d3dd9-3542-4351-8ca8-a55c7707e363 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.419275] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Doing hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 757.424168] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0a7d762d-2f4a-497b-a2cc-4a25cf112959 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.426463] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155923, 'name': ReconfigVM_Task, 'duration_secs': 0.550723} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.426852] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 757.426852] env[62066]: value = "task-1155927" [ 757.426852] env[62066]: _type = "Task" [ 757.426852] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.427261] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 248f5281-eb46-4be1-8642-28813c4b2622/248f5281-eb46-4be1-8642-28813c4b2622.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.428494] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83a0dc05-31ca-4d21-a304-82d31f91e722 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.436082] env[62066]: DEBUG oslo_concurrency.lockutils [req-eaebf34b-2502-487f-a00d-080152acdad0 req-270261d6-ee85-4d11-95a8-ea6dbb9cef86 service nova] Releasing lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.436705] env[62066]: DEBUG oslo_vmware.api [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 757.436705] env[62066]: value = "task-1155928" [ 757.436705] env[62066]: _type = "Task" [ 757.436705] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.453581] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155927, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.453901] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 757.453901] env[62066]: value = "task-1155929" [ 757.453901] env[62066]: _type = "Task" [ 757.453901] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.465690] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155926, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.465970] env[62066]: DEBUG oslo_vmware.api [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155928, 'name': ResetVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.471634] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155929, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.474293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "refresh_cache-de9eb96d-e71b-4878-83b8-f75966dc3f48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.513910] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5292588a-23c9-74ba-03e5-bc807f156a2c, 'name': SearchDatastore_Task, 'duration_secs': 0.019455} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.514300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.514536] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.514893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.547761] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524b66d7-a7f8-ea6a-b26f-a8999416ec50, 'name': SearchDatastore_Task, 'duration_secs': 0.010015} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.548070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.548451] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 5db50c22-048b-4cce-962a-3df1262f6e4f/5db50c22-048b-4cce-962a-3df1262f6e4f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.548693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.548884] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.549119] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d20b2772-fc02-46f3-9dc8-b05987fb3bec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.551301] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d39ea34-7105-4f1f-be45-51cb50460e5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.557703] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 757.557703] env[62066]: value = "task-1155930" [ 757.557703] env[62066]: _type = "Task" [ 757.557703] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.561997] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.562201] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.563294] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3e2440f-7e3d-45e9-9f6b-b6e4622c92a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.570106] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.571610] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 757.571610] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5222a8b8-fdb0-cd90-016b-413c29d14229" [ 757.571610] env[62066]: _type = "Task" [ 757.571610] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.580453] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5222a8b8-fdb0-cd90-016b-413c29d14229, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.843687] env[62066]: DEBUG nova.compute.utils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 757.851811] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.852027] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.941089] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155927, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.958154] env[62066]: DEBUG oslo_vmware.api [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155928, 'name': ResetVM_Task, 'duration_secs': 0.104046} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.967215] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Did hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 757.968603] env[62066]: DEBUG nova.compute.manager [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.968767] env[62066]: DEBUG oslo_vmware.api [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1155926, 'name': PowerOnVM_Task, 'duration_secs': 0.645244} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.978164] env[62066]: DEBUG nova.policy [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0bd92c6fd05488e9bf8f95b97077bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e781d81f23142d4859c188db11c59b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.978164] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caf9bd3-f290-4a71-9d0b-36bc2a6ed505 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.988340] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.990033] env[62066]: INFO nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Took 9.27 seconds to spawn the instance on the hypervisor. [ 757.990033] env[62066]: DEBUG nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.996459] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3d64f0-bbe4-4cab-94d3-e69848beb210 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.012331] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155929, 'name': Rename_Task, 'duration_secs': 0.34697} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.017193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.020392] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.022338] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a11f1c61-1c79-4688-9a28-fb9a60886b66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.024133] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c27cc9cd-e513-4ef3-b94b-26a075adbd17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.041173] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 758.041173] env[62066]: value = "task-1155931" [ 758.041173] env[62066]: _type = "Task" [ 758.041173] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.042634] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 758.042634] env[62066]: value = "task-1155932" [ 758.042634] env[62066]: _type = "Task" [ 758.042634] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.058657] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155931, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.062892] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.079527] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155930, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.088498] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5222a8b8-fdb0-cd90-016b-413c29d14229, 'name': SearchDatastore_Task, 'duration_secs': 0.011652} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.088910] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab30941f-66f0-44f1-959b-47d2e615e17d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.100958] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 758.100958] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d21da4-a03a-d14e-647c-b086c510e10a" [ 758.100958] env[62066]: _type = "Task" [ 758.100958] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.113758] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d21da4-a03a-d14e-647c-b086c510e10a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.206343] env[62066]: DEBUG nova.network.neutron [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Updated VIF entry in instance network info cache for port 2998edab-7365-4088-a93e-99cbbd0456fe. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 758.206735] env[62066]: DEBUG nova.network.neutron [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Updating instance_info_cache with network_info: [{"id": "2998edab-7365-4088-a93e-99cbbd0456fe", "address": "fa:16:3e:21:2e:c5", "network": {"id": "e3c5de25-e78b-49d2-b3c8-3f321eae6068", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1465671414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db9c97ec5490499bb671cbbae656b3fd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2998edab-73", "ovs_interfaceid": "2998edab-7365-4088-a93e-99cbbd0456fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.360894] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.394521] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad41a075-fe6d-41c4-aff9-a883eab256b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.403698] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043ccfff-5295-40b8-ba7b-ac5f62e7ffcc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.439407] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33e3149-4553-44c8-80b2-424fe05c6b65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.449138] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155927, 'name': Destroy_Task, 'duration_secs': 0.763467} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.451347] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Destroyed the VM [ 758.451598] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 758.452042] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-399b4310-2a4d-44f7-9262-0b5cc5cb8686 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.454534] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c6b49c-a438-49ce-89e2-57d300cb4072 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.469506] env[62066]: DEBUG nova.compute.provider_tree [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.471780] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 758.471780] env[62066]: value = "task-1155933" [ 758.471780] env[62066]: _type = "Task" [ 758.471780] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.479374] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155933, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.533096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-da9df6a8-1960-43ee-8d14-b92376ee8017 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.430s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.550934] env[62066]: INFO nova.compute.manager [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Took 32.52 seconds to build instance. [ 758.560892] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155931, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.564509] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155932, 'name': PowerOffVM_Task, 'duration_secs': 0.406337} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.567772] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.568732] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf4d333-28f1-4678-a50b-4f1562f412d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.575970] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155930, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760399} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.588981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 5db50c22-048b-4cce-962a-3df1262f6e4f/5db50c22-048b-4cce-962a-3df1262f6e4f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.589194] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.589879] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54bf129e-e9aa-4201-a8a0-6e9e6c23029a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.592418] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f10ec1-f037-4a50-98b5-7539ef87f2a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.604871] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 758.604871] env[62066]: value = "task-1155934" [ 758.604871] env[62066]: _type = "Task" [ 758.604871] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.614352] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d21da4-a03a-d14e-647c-b086c510e10a, 'name': SearchDatastore_Task, 'duration_secs': 0.053827} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.617636] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.617937] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 1c694323-616c-43b1-bd49-bba707e0788f/1c694323-616c-43b1-bd49-bba707e0788f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.620973] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155934, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.620973] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-468a4d17-4ab5-42ce-bf4e-7db0152bb676 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.625443] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 758.625443] env[62066]: value = "task-1155935" [ 758.625443] env[62066]: _type = "Task" [ 758.625443] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.636328] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.638742] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.639541] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-944cbd51-192b-47a8-adb5-b80ef8418f0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.646188] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 758.646188] env[62066]: value = "task-1155936" [ 758.646188] env[62066]: _type = "Task" [ 758.646188] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.656726] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 758.656973] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.657246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.657394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.657604] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.657874] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc371d6f-a189-47ce-9f69-c5bd8bd15fc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.667087] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.667291] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.668030] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cc694f6-1172-4b94-8c46-4910b0648f69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.673051] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 758.673051] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f0f3f9-6b84-de6a-8e55-6e555f0a0cd1" [ 758.673051] env[62066]: _type = "Task" [ 758.673051] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.681163] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f0f3f9-6b84-de6a-8e55-6e555f0a0cd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.711666] env[62066]: DEBUG oslo_concurrency.lockutils [req-890f7e36-81d2-43a3-82b6-d21a34e75ca5 req-8386a0bb-5459-4458-aa4c-eef1bb4d3e19 service nova] Releasing lock "refresh_cache-1c694323-616c-43b1-bd49-bba707e0788f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.720720] env[62066]: DEBUG nova.compute.manager [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.720720] env[62066]: DEBUG nova.compute.manager [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing instance network info cache due to event network-changed-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 758.720720] env[62066]: DEBUG oslo_concurrency.lockutils [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] Acquiring lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.720720] env[62066]: DEBUG oslo_concurrency.lockutils [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] Acquired lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.720720] env[62066]: DEBUG nova.network.neutron [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Refreshing network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 758.776557] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Successfully created port: 4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.974107] env[62066]: DEBUG nova.scheduler.client.report [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.989080] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155933, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.049968] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Successfully updated port: 600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 759.062223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-be984140-95dd-4e2f-a99b-c517c7239b00 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.689s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.063445] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155931, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.121891] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155934, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069444} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.122744] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.124156] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794a7700-eb35-48e1-8e14-5d2cfa9b04c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.165570] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 5db50c22-048b-4cce-962a-3df1262f6e4f/5db50c22-048b-4cce-962a-3df1262f6e4f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.171934] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-441070d9-195a-4e06-900b-a1b5f0f1abaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.192185] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155935, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.204040] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f0f3f9-6b84-de6a-8e55-6e555f0a0cd1, 'name': SearchDatastore_Task, 'duration_secs': 0.030679} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.207215] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 759.207215] env[62066]: value = "task-1155937" [ 759.207215] env[62066]: _type = "Task" [ 759.207215] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.207521] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e46586d-b5d8-47c2-9067-eb7336cf77c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.216983] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 759.216983] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521c8e99-a419-2ce3-2556-086f6a4d2f0f" [ 759.216983] env[62066]: _type = "Task" [ 759.216983] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.220861] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.234357] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521c8e99-a419-2ce3-2556-086f6a4d2f0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.380790] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.419038] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.419513] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.419715] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.419832] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.420052] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.420337] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.420411] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.420574] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.420907] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.421289] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.421588] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.423071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4a6fe4-3242-4dd2-aa64-339e0eb76aa6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.435475] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1f95c6-e5a5-4b31-9241-8bb3f9ab8a74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.483902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.145s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.486489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.213s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.494167] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155933, 'name': RemoveSnapshot_Task} progress is 29%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.514109] env[62066]: INFO nova.scheduler.client.report [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleted allocations for instance 265906be-704e-4f0f-ae33-e89135e27268 [ 759.556384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.556384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquired lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.556384] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.558135] env[62066]: DEBUG oslo_vmware.api [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155931, 'name': PowerOnVM_Task, 'duration_secs': 1.090567} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.558405] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.558635] env[62066]: INFO nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Took 13.33 seconds to spawn the instance on the hypervisor. [ 759.558809] env[62066]: DEBUG nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 759.559910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd51aa3-f47a-451d-9266-6b257934d6a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.567188] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.639389] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.836744} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.639696] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 1c694323-616c-43b1-bd49-bba707e0788f/1c694323-616c-43b1-bd49-bba707e0788f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.639934] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.640214] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38c7236b-633c-475a-8c42-6bae4ad70d8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.647802] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 759.647802] env[62066]: value = "task-1155938" [ 759.647802] env[62066]: _type = "Task" [ 759.647802] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.660030] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.694864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "641bf537-76d2-4c41-a6dc-b5b042327477" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.695383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.695484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.695691] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.695872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.699973] env[62066]: INFO nova.compute.manager [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Terminating instance [ 759.701955] env[62066]: DEBUG nova.compute.manager [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.702187] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.703061] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8323fe49-d033-4146-a37f-98f49a4cb572 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.711639] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 759.715281] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99991851-40d9-4ed2-ace7-f2b3dd033ffc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.722865] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.727658] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 759.727658] env[62066]: value = "task-1155939" [ 759.727658] env[62066]: _type = "Task" [ 759.727658] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.734937] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521c8e99-a419-2ce3-2556-086f6a4d2f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.025174} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.735681] env[62066]: DEBUG oslo_concurrency.lockutils [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.736012] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 759.736312] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a22a833d-0962-4e65-901b-f3717c5b1aca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.741719] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.746781] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 759.746781] env[62066]: value = "task-1155940" [ 759.746781] env[62066]: _type = "Task" [ 759.746781] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.757325] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155940, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.989280] env[62066]: DEBUG oslo_vmware.api [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155933, 'name': RemoveSnapshot_Task, 'duration_secs': 1.494723} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.989731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 759.989896] env[62066]: INFO nova.compute.manager [None req-153bacce-c32b-408b-951d-aedf490d1367 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Took 18.17 seconds to snapshot the instance on the hypervisor. [ 760.024651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d4249962-20a2-49e1-bc31-5c0cca8fb8ba tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "265906be-704e-4f0f-ae33-e89135e27268" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.592s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.093144] env[62066]: INFO nova.compute.manager [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Took 36.74 seconds to build instance. [ 760.096449] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.116225] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.158097] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071526} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.158389] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.159184] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e2c4c4-3b7d-4e8b-8fc7-5be2b5102bb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.184712] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 1c694323-616c-43b1-bd49-bba707e0788f/1c694323-616c-43b1-bd49-bba707e0788f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.185140] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bf6483c-09e8-48ee-8ee4-4f33de79f345 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.206273] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 760.206273] env[62066]: value = "task-1155941" [ 760.206273] env[62066]: _type = "Task" [ 760.206273] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.215744] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155941, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.228013] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155937, 'name': ReconfigVM_Task, 'duration_secs': 0.640941} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.228880] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 5db50c22-048b-4cce-962a-3df1262f6e4f/5db50c22-048b-4cce-962a-3df1262f6e4f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.232492] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95670539-823c-4e04-bea6-e8e064edf4c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.236113] env[62066]: DEBUG nova.network.neutron [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updated VIF entry in instance network info cache for port 5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.236421] env[62066]: DEBUG nova.network.neutron [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [{"id": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "address": "fa:16:3e:c0:80:33", "network": {"id": "b7dda1c4-dfff-4d56-b823-34efe1a5e870", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-488090776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "120bc7d4eb004ee3982fad8283f6f885", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8ba1e2-a1", "ovs_interfaceid": "5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.246743] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155939, 'name': PowerOffVM_Task, 'duration_secs': 0.329922} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.249701] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 760.249701] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 760.249701] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 760.249701] env[62066]: value = "task-1155942" [ 760.249701] env[62066]: _type = "Task" [ 760.249701] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.250791] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e3be14f-50f6-40a4-bee1-09db5813431b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.270340] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155942, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.270569] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155940, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.305527] env[62066]: DEBUG nova.network.neutron [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Updating instance_info_cache with network_info: [{"id": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "address": "fa:16:3e:b1:bb:74", "network": {"id": "4522831c-05ec-4d7a-b80b-5321d8f85fe2", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-643579621-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb0ca5356fdc4cbca29f97315a25483b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96fdbb91-eb49-4dbf-b234-5b38503d7589", "external-id": "nsx-vlan-transportzone-392", "segmentation_id": 392, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600db1e2-96", "ovs_interfaceid": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.338137] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 760.338447] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 760.338658] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Deleting the datastore file [datastore2] 641bf537-76d2-4c41-a6dc-b5b042327477 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.338940] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ea48960-84e7-4aed-8046-3f5bcacff2a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.346825] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for the task: (returnval){ [ 760.346825] env[62066]: value = "task-1155944" [ 760.346825] env[62066]: _type = "Task" [ 760.346825] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.358747] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155944, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.505413] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Applying migration context for instance c0ac4362-766f-48ba-aeb2-7fd976c1f47f as it has an incoming, in-progress migration de64f433-6637-4e88-8678-c58a6683456c. Migration status is reverting {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 760.507776] env[62066]: INFO nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating resource usage from migration de64f433-6637-4e88-8678-c58a6683456c [ 760.534454] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.534607] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.534731] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2f186d0a-91a0-4dc2-83bd-511099445af7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.534852] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.534980] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d50f0505-66f4-412f-9744-25e0ea96277a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535114] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a8dd7483-0588-4f60-9504-20de799e69f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535229] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535397] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6a498481-b2ca-4813-87b7-2f09dfa107f4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 760.535521] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535635] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 641bf537-76d2-4c41-a6dc-b5b042327477 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535748] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535859] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance de9eb96d-e71b-4878-83b8-f75966dc3f48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.535986] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 248f5281-eb46-4be1-8642-28813c4b2622 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.536088] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a2251e56-2787-412d-89c9-eef111ee6d2b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.536198] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.536306] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 1c694323-616c-43b1-bd49-bba707e0788f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.536412] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3d001331-7a41-4385-a622-54ce481509a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.536516] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 92b150af-351a-43ab-952a-209225733ac8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 760.598474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9254c50a-43d9-4179-90a5-df8de97024f5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.126s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.718982] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.739817] env[62066]: DEBUG oslo_concurrency.lockutils [req-8be47a37-e231-4187-b800-3757ec0f4a37 req-3ffa8e07-02ab-4e05-8fe9-95b678426bab service nova] Releasing lock "refresh_cache-641bf537-76d2-4c41-a6dc-b5b042327477" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.760356] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155940, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.945326} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.763552] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 760.764397] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f462138-4146-478c-8875-84940dd38590 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.771928] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155942, 'name': Rename_Task, 'duration_secs': 0.272992} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.786034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 760.793506] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.793818] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdfb5e39-3de1-4512-9f62-dddafb9c7418 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.795534] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daae6279-52dd-4acc-8d89-1d80dd2fda08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.808177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Releasing lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.808481] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Instance network_info: |[{"id": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "address": "fa:16:3e:b1:bb:74", "network": {"id": "4522831c-05ec-4d7a-b80b-5321d8f85fe2", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-643579621-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb0ca5356fdc4cbca29f97315a25483b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96fdbb91-eb49-4dbf-b234-5b38503d7589", "external-id": "nsx-vlan-transportzone-392", "segmentation_id": 392, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600db1e2-96", "ovs_interfaceid": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 760.809310] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:bb:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '96fdbb91-eb49-4dbf-b234-5b38503d7589', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '600db1e2-9642-4dd5-915b-ba68a16c87f4', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.816649] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Creating folder: Project (bb0ca5356fdc4cbca29f97315a25483b). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.818450] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4a1940b-733a-4ddf-8844-d05e8e6e5659 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.820505] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 760.820505] env[62066]: value = "task-1155945" [ 760.820505] env[62066]: _type = "Task" [ 760.820505] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.821845] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 760.821845] env[62066]: value = "task-1155946" [ 760.821845] env[62066]: _type = "Task" [ 760.821845] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.829592] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Created folder: Project (bb0ca5356fdc4cbca29f97315a25483b) in parent group-v251573. [ 760.830044] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Creating folder: Instances. Parent ref: group-v251650. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.833175] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca80a0aa-0504-42b0-beb4-5f621dc65617 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.834640] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155945, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.837724] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155946, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.846436] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Created folder: Instances in parent group-v251650. [ 760.846713] env[62066]: DEBUG oslo.service.loopingcall [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.846923] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 760.847252] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f9faab4-a88c-449b-ad0a-725ff6caa5b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.872064] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155944, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.873316] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.873316] env[62066]: value = "task-1155949" [ 760.873316] env[62066]: _type = "Task" [ 760.873316] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.881212] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155949, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.039996] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance c6fad95f-9167-4ad3-9c1c-16425fa1d36c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.099310] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.219201] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155941, 'name': ReconfigVM_Task, 'duration_secs': 0.734311} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.219667] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 1c694323-616c-43b1-bd49-bba707e0788f/1c694323-616c-43b1-bd49-bba707e0788f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.220256] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ace7900b-b3bb-467f-922c-1100651dab06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.230239] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 761.230239] env[62066]: value = "task-1155950" [ 761.230239] env[62066]: _type = "Task" [ 761.230239] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.241433] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155950, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.337955] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155945, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.342800] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155946, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.356720] env[62066]: DEBUG oslo_vmware.api [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Task: {'id': task-1155944, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.60268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.357084] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.357325] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.357524] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.357830] env[62066]: INFO nova.compute.manager [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Took 1.66 seconds to destroy the instance on the hypervisor. [ 761.358170] env[62066]: DEBUG oslo.service.loopingcall [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.358360] env[62066]: DEBUG nova.compute.manager [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.358483] env[62066]: DEBUG nova.network.neutron [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.369932] env[62066]: DEBUG nova.compute.manager [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Received event network-vif-plugged-600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.370106] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Acquiring lock "3d001331-7a41-4385-a622-54ce481509a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.370230] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Lock "3d001331-7a41-4385-a622-54ce481509a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.370573] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Lock "3d001331-7a41-4385-a622-54ce481509a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.370922] env[62066]: DEBUG nova.compute.manager [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] No waiting events found dispatching network-vif-plugged-600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 761.370922] env[62066]: WARNING nova.compute.manager [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Received unexpected event network-vif-plugged-600db1e2-9642-4dd5-915b-ba68a16c87f4 for instance with vm_state building and task_state spawning. [ 761.371060] env[62066]: DEBUG nova.compute.manager [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Received event network-changed-600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.371253] env[62066]: DEBUG nova.compute.manager [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Refreshing instance network info cache due to event network-changed-600db1e2-9642-4dd5-915b-ba68a16c87f4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.371456] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Acquiring lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.371811] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Acquired lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.371958] env[62066]: DEBUG nova.network.neutron [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Refreshing network info cache for port 600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.390841] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155949, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.440484] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Successfully updated port: 4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 761.542619] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d9163e64-309b-4381-8819-15757f83ac2e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.629424] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.742615] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155950, 'name': Rename_Task, 'duration_secs': 0.285919} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.743024] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.743305] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff9db6bf-0b15-4cd6-81c1-09d3e2da7e44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.749892] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 761.749892] env[62066]: value = "task-1155951" [ 761.749892] env[62066]: _type = "Task" [ 761.749892] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.760053] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155951, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.812762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c90e854-cb51-48ce-b097-acb29466e1b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.834534] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 761.838662] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-af2f251d-016f-4465-a72b-a7d95592657d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.844279] env[62066]: DEBUG oslo_vmware.api [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1155945, 'name': PowerOnVM_Task, 'duration_secs': 0.896793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.849020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 761.849020] env[62066]: INFO nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Took 10.42 seconds to spawn the instance on the hypervisor. [ 761.849020] env[62066]: DEBUG nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 761.851836] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c8c887-250b-4693-a142-d45fe32e5765 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.854672] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155946, 'name': ReconfigVM_Task, 'duration_secs': 0.662626} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.859057] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Reconfigured VM instance instance-00000035 to attach disk [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.859517] env[62066]: DEBUG oslo_vmware.api [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 761.859517] env[62066]: value = "task-1155952" [ 761.859517] env[62066]: _type = "Task" [ 761.859517] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.860911] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0503f5ef-19c6-411c-abd6-16876d3544b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.911788] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.912129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.912413] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.913423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.913423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.914747] env[62066]: DEBUG oslo_vmware.api [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155952, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.920370] env[62066]: INFO nova.compute.manager [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Terminating instance [ 761.923395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dace88d-889a-42d9-bdcc-710af9998a81 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.937725] env[62066]: DEBUG nova.compute.manager [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.939032] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.939032] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155949, 'name': CreateVM_Task, 'duration_secs': 0.573812} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.939226] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cd63d0-ab0a-4702-bfdb-d468272dce32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.943024] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.943748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.943898] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.944055] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.945910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.945910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.946126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.948130] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9185d03b-1a3b-42f9-83df-b4f8cafa2a91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.950248] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 761.950248] env[62066]: value = "task-1155953" [ 761.950248] env[62066]: _type = "Task" [ 761.950248] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.953715] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 761.959584] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41917798-edd8-42d2-be10-f1ce082c6380 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.964962] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 761.964962] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526656b2-99cd-6d1e-dfad-4edfbcf34019" [ 761.964962] env[62066]: _type = "Task" [ 761.964962] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.979233] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.979980] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 761.979980] env[62066]: value = "task-1155954" [ 761.979980] env[62066]: _type = "Task" [ 761.979980] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.989214] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526656b2-99cd-6d1e-dfad-4edfbcf34019, 'name': SearchDatastore_Task, 'duration_secs': 0.011339} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.990304] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.990452] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.990783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.991049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.991307] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.995270] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c6ba9d8-4d40-4e87-bc60-78b2321f88cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.998217] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.005050] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.005300] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.006481] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d4774ae-79b0-42cd-8707-57c5e8f42911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.012714] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 762.012714] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52faa067-5ac6-cc85-9ce1-b7ee2fc962cf" [ 762.012714] env[62066]: _type = "Task" [ 762.012714] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.025354] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52faa067-5ac6-cc85-9ce1-b7ee2fc962cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.046363] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.046879] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Migration de64f433-6637-4e88-8678-c58a6683456c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 762.148105] env[62066]: DEBUG nova.network.neutron [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.268291] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155951, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.268291] env[62066]: DEBUG nova.compute.manager [req-c77a61fa-6278-44f1-a83f-b4dcfd24daae req-476dc65d-b3eb-44c7-86a1-c42f574228cd service nova] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Received event network-vif-deleted-5b8ba1e2-a1bc-4e81-9ff7-07c21d260c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.371272] env[62066]: DEBUG nova.network.neutron [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Updated VIF entry in instance network info cache for port 600db1e2-9642-4dd5-915b-ba68a16c87f4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 762.371692] env[62066]: DEBUG nova.network.neutron [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Updating instance_info_cache with network_info: [{"id": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "address": "fa:16:3e:b1:bb:74", "network": {"id": "4522831c-05ec-4d7a-b80b-5321d8f85fe2", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-643579621-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bb0ca5356fdc4cbca29f97315a25483b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96fdbb91-eb49-4dbf-b234-5b38503d7589", "external-id": "nsx-vlan-transportzone-392", "segmentation_id": 392, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap600db1e2-96", "ovs_interfaceid": "600db1e2-9642-4dd5-915b-ba68a16c87f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.380905] env[62066]: DEBUG oslo_vmware.api [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155952, 'name': SuspendVM_Task} progress is 50%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.387791] env[62066]: INFO nova.compute.manager [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Took 32.03 seconds to build instance. [ 762.471060] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155953, 'name': ReconfigVM_Task, 'duration_secs': 0.207911} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.471381] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.471641] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0aadd556-de3a-4624-98be-7f27b70303ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.479813] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 762.479813] env[62066]: value = "task-1155955" [ 762.479813] env[62066]: _type = "Task" [ 762.479813] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.491720] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.494997] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155954, 'name': PowerOffVM_Task, 'duration_secs': 0.204184} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.495442] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.495519] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.495806] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66507c0d-f5d6-4554-91d3-bb6db4faa388 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.526827] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52faa067-5ac6-cc85-9ce1-b7ee2fc962cf, 'name': SearchDatastore_Task, 'duration_secs': 0.013577} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.528915] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ad1ff01-ea84-48af-9d45-12c5602d884f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.534784] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 762.534784] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52da652a-4ad5-9554-0e86-ab08db0daba1" [ 762.534784] env[62066]: _type = "Task" [ 762.534784] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.545423] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52da652a-4ad5-9554-0e86-ab08db0daba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.549405] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance e95ee795-603e-4cbf-bcd6-1ba54b62a281 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.549625] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance c0ac4362-766f-48ba-aeb2-7fd976c1f47f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 762.562980] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.563261] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.563451] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Deleting the datastore file [datastore1] 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.564586] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-368573b4-dbb5-461e-9ed9-0948e3218d56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.571196] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for the task: (returnval){ [ 762.571196] env[62066]: value = "task-1155957" [ 762.571196] env[62066]: _type = "Task" [ 762.571196] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.580481] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.634772] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.654404] env[62066]: INFO nova.compute.manager [-] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Took 1.29 seconds to deallocate network for instance. [ 762.761512] env[62066]: DEBUG oslo_vmware.api [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155951, 'name': PowerOnVM_Task, 'duration_secs': 0.75787} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.761791] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.762014] env[62066]: INFO nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Took 8.75 seconds to spawn the instance on the hypervisor. [ 762.762209] env[62066]: DEBUG nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.763009] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25eb16b3-b258-48f3-8ac8-cd4fa3452744 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.833289] env[62066]: DEBUG nova.compute.manager [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.834199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4f9028-e6e1-44ef-8b2b-f4dba0dc6536 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.883644] env[62066]: DEBUG oslo_concurrency.lockutils [req-bbe81081-32f5-4512-b284-9befd99398c1 req-4eb823fa-65f8-4114-a715-bf00a5f93268 service nova] Releasing lock "refresh_cache-3d001331-7a41-4385-a622-54ce481509a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.884608] env[62066]: DEBUG oslo_vmware.api [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155952, 'name': SuspendVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.892521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e69bc7da-d8ef-4eb6-b8ea-43c7acacc57a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.847s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.928132] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Updating instance_info_cache with network_info: [{"id": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "address": "fa:16:3e:c2:02:ff", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7a82c6-2b", "ovs_interfaceid": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.991022] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155955, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.050178] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52da652a-4ad5-9554-0e86-ab08db0daba1, 'name': SearchDatastore_Task, 'duration_secs': 0.019835} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.050787] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.051101] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3d001331-7a41-4385-a622-54ce481509a8/3d001331-7a41-4385-a622-54ce481509a8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.051395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f28c0132-2725-4fbd-86ad-44242a9ebaa3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.054339] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 43d64f76-7f7a-4b95-b9df-c95218612998 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.060833] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 763.060833] env[62066]: value = "task-1155958" [ 763.060833] env[62066]: _type = "Task" [ 763.060833] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.070513] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155958, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.082596] env[62066]: DEBUG oslo_vmware.api [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Task: {'id': task-1155957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329651} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.082815] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.083051] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.083504] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.083734] env[62066]: INFO nova.compute.manager [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 763.084064] env[62066]: DEBUG oslo.service.loopingcall [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.084313] env[62066]: DEBUG nova.compute.manager [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.084446] env[62066]: DEBUG nova.network.neutron [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.163897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.280037] env[62066]: INFO nova.compute.manager [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Took 27.56 seconds to build instance. [ 763.348361] env[62066]: INFO nova.compute.manager [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] instance snapshotting [ 763.351942] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9407d2-98e8-43fd-8123-36ab0b8c1faa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.379049] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b751032-1444-4931-a082-cb668290f73b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.387461] env[62066]: DEBUG oslo_vmware.api [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155952, 'name': SuspendVM_Task, 'duration_secs': 1.073321} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.389389] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 763.389528] env[62066]: DEBUG nova.compute.manager [None req-15d45b75-8697-4861-812c-9f50f00177e5 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 763.392875] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c98fa38-bdae-4985-bf39-4b01c0c14409 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.395659] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.432685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.432685] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Instance network_info: |[{"id": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "address": "fa:16:3e:c2:02:ff", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7a82c6-2b", "ovs_interfaceid": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 763.432685] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:02:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6e940e5-e083-4238-973e-f1b4e2a3a5c7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f7a82c6-2b47-414a-9821-e48187f88ba3', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.446584] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Creating folder: Project (8e781d81f23142d4859c188db11c59b0). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.447489] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-646bca6a-80ad-4f67-9a27-408b525a372f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.457992] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Created folder: Project (8e781d81f23142d4859c188db11c59b0) in parent group-v251573. [ 763.458575] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Creating folder: Instances. Parent ref: group-v251653. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 763.458881] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d43ebc3-a6f5-461a-8aa1-cdad2bea3228 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.472468] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Created folder: Instances in parent group-v251653. [ 763.472733] env[62066]: DEBUG oslo.service.loopingcall [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.472931] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.476206] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06cd294b-f1ba-4ef3-8d97-2d4590aa5ae4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.503197] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.503197] env[62066]: value = "task-1155961" [ 763.503197] env[62066]: _type = "Task" [ 763.503197] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.512739] env[62066]: DEBUG oslo_vmware.api [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1155955, 'name': PowerOnVM_Task, 'duration_secs': 0.63714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.515225] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.521574] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155961, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.522803] env[62066]: DEBUG nova.compute.manager [None req-248d5527-9823-4af3-9138-226b93fcbc49 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 763.524029] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e374c9fa-3e6c-4edf-9812-915a666d3b02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.557705] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 4723be94-3479-4e66-8088-914824c0e669 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 763.572234] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155958, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.726838] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.727161] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing instance network info cache due to event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 763.728183] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.728183] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.728355] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 763.782371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0324912-62bf-4c6f-84d0-1e5dbeb7bd19 tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.742s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.897440] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 763.897785] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a8091444-dfa4-4d90-a9a1-0b63650fbda0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.908050] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 763.908050] env[62066]: value = "task-1155962" [ 763.908050] env[62066]: _type = "Task" [ 763.908050] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.921194] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155962, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.932911] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.015152] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155961, 'name': CreateVM_Task, 'duration_secs': 0.431945} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.015798] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 764.017385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.017535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.018126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 764.018415] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa42f335-1781-4b25-9b06-daad84997c8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.023541] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 764.023541] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52afa73b-aea0-4578-ec23-f3d87a05709f" [ 764.023541] env[62066]: _type = "Task" [ 764.023541] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.033378] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52afa73b-aea0-4578-ec23-f3d87a05709f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.058241] env[62066]: DEBUG nova.network.neutron [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.062201] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance b46b7795-b7f7-4c55-bc72-184bb7bb8842 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 764.062201] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 764.062201] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4224MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 764.079391] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155958, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706342} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.079718] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3d001331-7a41-4385-a622-54ce481509a8/3d001331-7a41-4385-a622-54ce481509a8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 764.080035] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.080909] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-373a6eb5-a1b0-423b-8ec2-ed289c597206 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.093151] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 764.093151] env[62066]: value = "task-1155963" [ 764.093151] env[62066]: _type = "Task" [ 764.093151] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.107538] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.294863] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.294863] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.420421] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155962, 'name': CreateSnapshot_Task, 'duration_secs': 0.479688} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.420753] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 764.421511] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505970f8-1bee-4b74-a026-1e5ad60be2b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.491492] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updated VIF entry in instance network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 764.491970] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.539201] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52afa73b-aea0-4578-ec23-f3d87a05709f, 'name': SearchDatastore_Task, 'duration_secs': 0.015774} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.539526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.539806] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.540073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.540291] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.540403] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.541548] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b5e7f07-d698-4746-b054-51cbf8cbc079 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.543966] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b500f4d-a6bf-4351-a8df-3a023411a04b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.553126] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f373b4a8-19db-4e94-a509-1340afd62109 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.557909] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.558096] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 764.559540] env[62066]: INFO nova.compute.manager [-] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Took 1.48 seconds to deallocate network for instance. [ 764.559738] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53b11c0c-5901-4cf1-b849-a838980c6d26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.592456] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59e1b8b-9935-4362-ba84-4c1e4f1ada03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.603598] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 764.603598] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cbe9b9-e468-dac8-bd43-c9266f511262" [ 764.603598] env[62066]: _type = "Task" [ 764.603598] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.614204] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148b83b7-484c-47f8-9c25-6be3269fdbcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.626199] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.626199] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cbe9b9-e468-dac8-bd43-c9266f511262, 'name': SearchDatastore_Task, 'duration_secs': 0.018276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.626443] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.627709] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c2e05e-dadd-40e3-b1de-b70e52b4d986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.630277] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a95f994-e333-42ed-a45c-a24a5ad7e34d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.642573] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.663492] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 3d001331-7a41-4385-a622-54ce481509a8/3d001331-7a41-4385-a622-54ce481509a8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.666186] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-516e7d36-8775-4f03-9613-8f90f3c2ea5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.681385] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 764.681385] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d21fa1-f5f8-da65-35f2-ac8d63874a8d" [ 764.681385] env[62066]: _type = "Task" [ 764.681385] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.687543] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 764.687543] env[62066]: value = "task-1155964" [ 764.687543] env[62066]: _type = "Task" [ 764.687543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.695587] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d21fa1-f5f8-da65-35f2-ac8d63874a8d, 'name': SearchDatastore_Task, 'duration_secs': 0.012021} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.699343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.699593] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 92b150af-351a-43ab-952a-209225733ac8/92b150af-351a-43ab-952a-209225733ac8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.699835] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155964, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.700382] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2530298b-5da8-47f6-a168-ae405b4abb33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.705644] env[62066]: DEBUG nova.compute.manager [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.705644] env[62066]: DEBUG nova.compute.manager [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing instance network info cache due to event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.705644] env[62066]: DEBUG oslo_concurrency.lockutils [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] Acquiring lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.705785] env[62066]: DEBUG oslo_concurrency.lockutils [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] Acquired lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.705990] env[62066]: DEBUG nova.network.neutron [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 764.710233] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 764.710233] env[62066]: value = "task-1155965" [ 764.710233] env[62066]: _type = "Task" [ 764.710233] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.718850] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.795938] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.941805] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 764.942799] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4f608ecb-4c84-4539-9b95-c88ecd9e101a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.951694] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 764.951694] env[62066]: value = "task-1155966" [ 764.951694] env[62066]: _type = "Task" [ 764.951694] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.961827] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155966, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.006026] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Received event network-vif-plugged-4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Acquiring lock "92b150af-351a-43ab-952a-209225733ac8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Lock "92b150af-351a-43ab-952a-209225733ac8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Lock "92b150af-351a-43ab-952a-209225733ac8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.006026] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] No waiting events found dispatching network-vif-plugged-4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 765.006026] env[62066]: WARNING nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Received unexpected event network-vif-plugged-4f7a82c6-2b47-414a-9821-e48187f88ba3 for instance with vm_state building and task_state spawning. [ 765.006026] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Received event network-changed-4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.006026] env[62066]: DEBUG nova.compute.manager [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Refreshing instance network info cache due to event network-changed-4f7a82c6-2b47-414a-9821-e48187f88ba3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Acquiring lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.006026] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Acquired lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.006026] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Refreshing network info cache for port 4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.105391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.146442] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.221670] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155964, 'name': ReconfigVM_Task, 'duration_secs': 0.446923} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.230086] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 3d001331-7a41-4385-a622-54ce481509a8/3d001331-7a41-4385-a622-54ce481509a8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.233896] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c2610c0-2cd4-4bb5-86df-72d6b96c4f83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.243044] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155965, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.243044] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 765.243044] env[62066]: value = "task-1155967" [ 765.243044] env[62066]: _type = "Task" [ 765.243044] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.255083] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155967, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.260211] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "f196adc6-86bb-4556-bf3d-c52b9108c000" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.260211] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.266050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "f3ef43e8-6092-44ac-8990-979810a9748f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.266050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.329263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.358633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "1c694323-616c-43b1-bd49-bba707e0788f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.359007] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.359541] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "1c694323-616c-43b1-bd49-bba707e0788f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.359794] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.360078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.363071] env[62066]: INFO nova.compute.manager [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Terminating instance [ 765.365388] env[62066]: DEBUG nova.compute.manager [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 765.365501] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.366375] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438a887a-6ef4-4a60-ab06-6aac747f35c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.375691] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 765.375691] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fd8826d-79ea-4d66-a5da-9a68446d4a5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.385034] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 765.385034] env[62066]: value = "task-1155968" [ 765.385034] env[62066]: _type = "Task" [ 765.385034] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.395830] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.462561] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155966, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.535907] env[62066]: DEBUG nova.network.neutron [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updated VIF entry in instance network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 765.536770] env[62066]: DEBUG nova.network.neutron [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.652527] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 765.652737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.166s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.654038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.360s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.654970] env[62066]: INFO nova.compute.claims [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.701641] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.733023] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558873} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.733023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 92b150af-351a-43ab-952a-209225733ac8/92b150af-351a-43ab-952a-209225733ac8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.733023] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.733023] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43f64cdd-99af-416d-a276-9dd2fada6e9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.738937] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 765.738937] env[62066]: value = "task-1155969" [ 765.738937] env[62066]: _type = "Task" [ 765.738937] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.759192] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155967, 'name': Rename_Task, 'duration_secs': 0.242536} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.759502] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.762927] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 765.764160] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1d2bcab-795f-4c83-9fcb-f72056f76eb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.773020] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 765.773020] env[62066]: value = "task-1155970" [ 765.773020] env[62066]: _type = "Task" [ 765.773020] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.780470] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.901334] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155968, 'name': PowerOffVM_Task, 'duration_secs': 0.213769} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.902045] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.902501] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.902998] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccb45633-ba4e-48a7-bfa8-d9abbc49102d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.950641] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Updated VIF entry in instance network info cache for port 4f7a82c6-2b47-414a-9821-e48187f88ba3. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 765.951062] env[62066]: DEBUG nova.network.neutron [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Updating instance_info_cache with network_info: [{"id": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "address": "fa:16:3e:c2:02:ff", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f7a82c6-2b", "ovs_interfaceid": "4f7a82c6-2b47-414a-9821-e48187f88ba3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.961646] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155966, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.985877] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.986145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.986352] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Deleting the datastore file [datastore1] 1c694323-616c-43b1-bd49-bba707e0788f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.986685] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ce9e8ed-4575-48df-8c95-56def1d3ea70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.993943] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for the task: (returnval){ [ 765.993943] env[62066]: value = "task-1155972" [ 765.993943] env[62066]: _type = "Task" [ 765.993943] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.003042] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.039176] env[62066]: DEBUG oslo_concurrency.lockutils [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] Releasing lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.039452] env[62066]: DEBUG nova.compute.manager [req-3fdd0bae-7eec-4270-bfb9-69e25936d9b1 req-74331f99-fccc-431c-89ee-47f119da22e0 service nova] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Received event network-vif-deleted-4e4bf2f1-d884-4bd4-91b5-8d1574133055 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.206671] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.250948] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150753} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.251359] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.252259] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17f0064-7af5-491c-bc5c-5f22adcac83c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.280965] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 92b150af-351a-43ab-952a-209225733ac8/92b150af-351a-43ab-952a-209225733ac8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.282118] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c085fb-a2e1-40af-afaf-48248dc8fdfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.310568] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155970, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.313240] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 766.313240] env[62066]: value = "task-1155973" [ 766.313240] env[62066]: _type = "Task" [ 766.313240] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.322319] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.456694] env[62066]: DEBUG oslo_concurrency.lockutils [req-ee3eb12e-4851-4105-b214-fc3706c5833a req-5b452638-b135-42a3-9de5-2b2b781daa81 service nova] Releasing lock "refresh_cache-92b150af-351a-43ab-952a-209225733ac8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.463058] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155966, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.504027] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.783451] env[62066]: DEBUG nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.784640] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8fe398-6d1c-4d02-b7e1-8db9bc6637df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.804639] env[62066]: DEBUG oslo_vmware.api [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155970, 'name': PowerOnVM_Task, 'duration_secs': 0.878102} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.807430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 766.807663] env[62066]: INFO nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Took 10.14 seconds to spawn the instance on the hypervisor. [ 766.807872] env[62066]: DEBUG nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.808909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15832fed-9175-4a2c-acce-3f1e3b05eacf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.831617] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155973, 'name': ReconfigVM_Task, 'duration_secs': 0.320728} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.831864] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 92b150af-351a-43ab-952a-209225733ac8/92b150af-351a-43ab-952a-209225733ac8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.832581] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc1f969e-de11-494a-9fcd-2fbebec5edf7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.839848] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 766.839848] env[62066]: value = "task-1155974" [ 766.839848] env[62066]: _type = "Task" [ 766.839848] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.848102] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155974, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.961835] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155966, 'name': CloneVM_Task, 'duration_secs': 1.564782} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.964251] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Created linked-clone VM from snapshot [ 766.965260] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c378db3d-8890-4120-974c-028293a2225d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.972595] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Uploading image de8b415d-4ae3-4740-a106-cbd973431622 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 767.005791] env[62066]: DEBUG oslo_vmware.api [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Task: {'id': task-1155972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.513309} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.008887] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 767.008887] env[62066]: value = "vm-251657" [ 767.008887] env[62066]: _type = "VirtualMachine" [ 767.008887] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 767.008887] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 767.008887] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 767.009485] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 767.009485] env[62066]: INFO nova.compute.manager [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Took 1.64 seconds to destroy the instance on the hypervisor. [ 767.009674] env[62066]: DEBUG oslo.service.loopingcall [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.009985] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8ef13e8d-1ed2-43d4-bd02-6e98f33c92cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.013980] env[62066]: DEBUG nova.compute.manager [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.014111] env[62066]: DEBUG nova.network.neutron [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 767.020781] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease: (returnval){ [ 767.020781] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a8aa23-5fd7-847e-9757-93c2b62c773d" [ 767.020781] env[62066]: _type = "HttpNfcLease" [ 767.020781] env[62066]: } obtained for exporting VM: (result){ [ 767.020781] env[62066]: value = "vm-251657" [ 767.020781] env[62066]: _type = "VirtualMachine" [ 767.020781] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 767.020781] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the lease: (returnval){ [ 767.020781] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a8aa23-5fd7-847e-9757-93c2b62c773d" [ 767.020781] env[62066]: _type = "HttpNfcLease" [ 767.020781] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 767.031533] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 767.031533] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a8aa23-5fd7-847e-9757-93c2b62c773d" [ 767.031533] env[62066]: _type = "HttpNfcLease" [ 767.031533] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 767.089783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cab1d93-9956-4e70-8a79-cd8081ac7452 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.098224] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddfad25-b908-4299-aa8c-3b6d95da5986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.137336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061c4ab3-6095-4cf1-8c21-9d3855fff077 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.147755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eec636e-a895-472d-9ebb-8d9717e89455 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.163857] env[62066]: DEBUG nova.compute.provider_tree [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.199648] env[62066]: DEBUG nova.compute.manager [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.200714] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc823dd-d47e-4a8a-a240-009f7b280ccd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.307276] env[62066]: INFO nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] instance snapshotting [ 767.307599] env[62066]: WARNING nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 767.311204] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c78c9d-bd5f-4c44-afbd-f996751efe6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.340007] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5196971a-4c5a-4d01-bd48-974d75c0354f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.343197] env[62066]: INFO nova.compute.manager [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Took 30.01 seconds to build instance. [ 767.365898] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155974, 'name': Rename_Task, 'duration_secs': 0.170527} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.365898] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.366427] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-789b2649-147d-4ee0-80f4-9983e255344b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.376258] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 767.376258] env[62066]: value = "task-1155976" [ 767.376258] env[62066]: _type = "Task" [ 767.376258] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.387097] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.523286] env[62066]: DEBUG nova.compute.manager [req-b6ba6843-daab-4dd7-973c-d2809a07cd4f req-c46b8e08-b3bd-464a-8a83-1c9e76b99afe service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Received event network-vif-deleted-2998edab-7365-4088-a93e-99cbbd0456fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.523348] env[62066]: INFO nova.compute.manager [req-b6ba6843-daab-4dd7-973c-d2809a07cd4f req-c46b8e08-b3bd-464a-8a83-1c9e76b99afe service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Neutron deleted interface 2998edab-7365-4088-a93e-99cbbd0456fe; detaching it from the instance and deleting it from the info cache [ 767.523525] env[62066]: DEBUG nova.network.neutron [req-b6ba6843-daab-4dd7-973c-d2809a07cd4f req-c46b8e08-b3bd-464a-8a83-1c9e76b99afe service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.533231] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 767.533231] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a8aa23-5fd7-847e-9757-93c2b62c773d" [ 767.533231] env[62066]: _type = "HttpNfcLease" [ 767.533231] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 767.533567] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 767.533567] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a8aa23-5fd7-847e-9757-93c2b62c773d" [ 767.533567] env[62066]: _type = "HttpNfcLease" [ 767.533567] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 767.534355] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c807f60-ea55-4c15-a66f-6c64b50dc6d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.543886] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 767.544089] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 767.642063] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-81b5ccee-b111-4043-8d31-06ce3cd476d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.666993] env[62066]: DEBUG nova.scheduler.client.report [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.718412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "3d001331-7a41-4385-a622-54ce481509a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.719065] env[62066]: INFO nova.compute.manager [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] instance snapshotting [ 767.719711] env[62066]: DEBUG nova.objects.instance [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.845641] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a3af15-a82d-4c3a-9264-ed247a43f71a tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.846995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.129s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.847240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "3d001331-7a41-4385-a622-54ce481509a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.847450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.847614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.857869] env[62066]: INFO nova.compute.manager [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Terminating instance [ 767.862199] env[62066]: DEBUG nova.compute.manager [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.862199] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 767.863600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c28eae4-2802-443f-9b0d-0ad418fcc691 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.870093] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 767.870540] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-08d1f546-88fd-480d-aaf3-c4e3f9ab5d7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.880847] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 767.884214] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbbe5c2c-801f-4a38-8b9b-e36420ddb347 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.888023] env[62066]: DEBUG nova.network.neutron [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.891936] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 767.891936] env[62066]: value = "task-1155977" [ 767.891936] env[62066]: _type = "Task" [ 767.891936] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.903027] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155976, 'name': PowerOnVM_Task, 'duration_secs': 0.46534} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.903174] env[62066]: DEBUG oslo_vmware.api [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 767.903174] env[62066]: value = "task-1155978" [ 767.903174] env[62066]: _type = "Task" [ 767.903174] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.904474] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.904737] env[62066]: INFO nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Took 8.52 seconds to spawn the instance on the hypervisor. [ 767.904936] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.908231] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d3fd04-fd44-4a2e-b386-72f6d3a8c218 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.918224] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155977, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.927560] env[62066]: DEBUG oslo_vmware.api [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.030235] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92864d27-67bd-4ec4-bad6-d82273d3000d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.041017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a917cf31-b9cc-4da1-b938-bc33715823a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.077980] env[62066]: DEBUG nova.compute.manager [req-b6ba6843-daab-4dd7-973c-d2809a07cd4f req-c46b8e08-b3bd-464a-8a83-1c9e76b99afe service nova] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Detach interface failed, port_id=2998edab-7365-4088-a93e-99cbbd0456fe, reason: Instance 1c694323-616c-43b1-bd49-bba707e0788f could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 768.178089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.178844] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.182385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.141s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.182600] env[62066]: DEBUG nova.objects.instance [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 768.227481] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845d1b73-a8da-4087-a5c2-0f1f79ce3d82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.250627] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5bf444-c9ea-4fa3-98cd-2f58f9bbb3d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.361988] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.390617] env[62066]: INFO nova.compute.manager [-] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Took 1.38 seconds to deallocate network for instance. [ 768.405469] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155977, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.415747] env[62066]: DEBUG oslo_vmware.api [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155978, 'name': PowerOffVM_Task, 'duration_secs': 0.234249} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.416157] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.416384] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.416813] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5fa31d4-7d3b-46f3-8384-0c20bd29aee9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.438423] env[62066]: INFO nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Took 30.36 seconds to build instance. [ 768.684259] env[62066]: DEBUG nova.compute.utils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 768.687259] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.687495] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.760119] env[62066]: DEBUG nova.policy [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0bd92c6fd05488e9bf8f95b97077bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e781d81f23142d4859c188db11c59b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.762820] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 768.763245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-078b3b85-7c9e-493b-a303-2326918b8f6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.772611] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 768.772611] env[62066]: value = "task-1155980" [ 768.772611] env[62066]: _type = "Task" [ 768.772611] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.782659] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155980, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.892835] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.893082] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.893267] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Deleting the datastore file [datastore1] 3d001331-7a41-4385-a622-54ce481509a8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.893882] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84185998-b415-47e9-b98c-f0a8b1cfcae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.898096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.904425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.911658] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155977, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.913073] env[62066]: DEBUG oslo_vmware.api [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for the task: (returnval){ [ 768.913073] env[62066]: value = "task-1155981" [ 768.913073] env[62066]: _type = "Task" [ 768.913073] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.943217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.741s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.101489] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Successfully created port: 88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.189485] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.203483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c441af3a-1033-4bf2-97de-3be6467569e5 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.207780] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.409s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.207780] env[62066]: INFO nova.compute.claims [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.284340] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155980, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.408715] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155977, 'name': CreateSnapshot_Task, 'duration_secs': 1.105121} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.409011] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 769.409874] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33df4a57-f9a5-4a8d-8c51-dfd1d327510e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.428995] env[62066]: DEBUG oslo_vmware.api [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Task: {'id': task-1155981, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193219} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.429366] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.429663] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.429869] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.430102] env[62066]: INFO nova.compute.manager [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Took 1.57 seconds to destroy the instance on the hypervisor. [ 769.430534] env[62066]: DEBUG oslo.service.loopingcall [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.430951] env[62066]: DEBUG nova.compute.manager [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.431031] env[62066]: DEBUG nova.network.neutron [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 769.446429] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.783752] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155980, 'name': CreateSnapshot_Task, 'duration_secs': 0.615736} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.785142] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 769.786339] env[62066]: DEBUG nova.compute.manager [req-5874fc09-b798-4dff-81ff-02d28fc1818e req-e27374de-6387-4dd0-af26-e7d3b8d2bfbf service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Received event network-vif-deleted-600db1e2-9642-4dd5-915b-ba68a16c87f4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.786535] env[62066]: INFO nova.compute.manager [req-5874fc09-b798-4dff-81ff-02d28fc1818e req-e27374de-6387-4dd0-af26-e7d3b8d2bfbf service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Neutron deleted interface 600db1e2-9642-4dd5-915b-ba68a16c87f4; detaching it from the instance and deleting it from the info cache [ 769.786720] env[62066]: DEBUG nova.network.neutron [req-5874fc09-b798-4dff-81ff-02d28fc1818e req-e27374de-6387-4dd0-af26-e7d3b8d2bfbf service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.788316] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4276d5b8-caf8-4d3a-a484-4e3bcfc54b6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.932545] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 769.933326] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e0dd55e1-12d3-43f1-9a3c-6d09eabb6bf7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.943631] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 769.943631] env[62066]: value = "task-1155982" [ 769.943631] env[62066]: _type = "Task" [ 769.943631] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.955065] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155982, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.974407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.200970] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.208382] env[62066]: DEBUG nova.network.neutron [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.231402] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.231701] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.231884] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.232135] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.232296] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.232496] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.232678] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.232865] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.233075] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.233249] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.233513] env[62066]: DEBUG nova.virt.hardware [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.235300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ffb967-bc51-451d-bdcb-722131d0a3b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.249736] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f30b56-a88c-4d8d-8422-db509ae85b9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.293542] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc1bf1d3-65d1-45df-888c-bbf11c54e0be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.308031] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 770.312040] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aa378a04-1c51-477d-80fe-2bccc2a3a786 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.319990] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907c5c85-1d01-4e0c-a619-71be19973550 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.339728] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 770.339728] env[62066]: value = "task-1155983" [ 770.339728] env[62066]: _type = "Task" [ 770.339728] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.366520] env[62066]: DEBUG nova.compute.manager [req-5874fc09-b798-4dff-81ff-02d28fc1818e req-e27374de-6387-4dd0-af26-e7d3b8d2bfbf service nova] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Detach interface failed, port_id=600db1e2-9642-4dd5-915b-ba68a16c87f4, reason: Instance 3d001331-7a41-4385-a622-54ce481509a8 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 770.374897] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155983, 'name': CloneVM_Task} progress is 12%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.458111] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155982, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.711193] env[62066]: INFO nova.compute.manager [-] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Took 1.28 seconds to deallocate network for instance. [ 770.717767] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11717c9c-d4a1-4d0d-930b-2931e74eff28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.726991] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45898fd7-6add-4f05-b88a-7f258c6e4124 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.761195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4689dd0c-190c-421f-b99d-539f23cd37f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.770999] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5771924-3578-4821-96ed-822741ff5337 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.791400] env[62066]: DEBUG nova.compute.provider_tree [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.850624] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155983, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.933727] env[62066]: DEBUG nova.compute.manager [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Received event network-vif-plugged-88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.933995] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] Acquiring lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.936188] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.936188] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.936188] env[62066]: DEBUG nova.compute.manager [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] No waiting events found dispatching network-vif-plugged-88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 770.936188] env[62066]: WARNING nova.compute.manager [req-1c45056d-adae-40d1-bbe7-ec99c64b1ff8 req-68cd66cc-0a8f-4a8d-a459-abd6037179a9 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Received unexpected event network-vif-plugged-88572837-2f5a-41c7-9c89-df220eebdc16 for instance with vm_state building and task_state spawning. [ 770.954890] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155982, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.026466] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Successfully updated port: 88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.222560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.295203] env[62066]: DEBUG nova.scheduler.client.report [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.350867] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155983, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.456134] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1155982, 'name': CloneVM_Task, 'duration_secs': 1.448762} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.456411] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Created linked-clone VM from snapshot [ 771.457198] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7b7a4b-900f-4f33-a1b7-b18cac75f87a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.464665] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Uploading image 74f0bdd6-1197-4371-a231-52e7fe03597c {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 771.484778] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 771.484778] env[62066]: value = "vm-251660" [ 771.484778] env[62066]: _type = "VirtualMachine" [ 771.484778] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 771.485090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0f76c14e-8e51-4c63-b81f-40c0fe50c43b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.492320] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease: (returnval){ [ 771.492320] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b52aed-9a56-02d0-a365-28d50deabb94" [ 771.492320] env[62066]: _type = "HttpNfcLease" [ 771.492320] env[62066]: } obtained for exporting VM: (result){ [ 771.492320] env[62066]: value = "vm-251660" [ 771.492320] env[62066]: _type = "VirtualMachine" [ 771.492320] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 771.492656] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the lease: (returnval){ [ 771.492656] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b52aed-9a56-02d0-a365-28d50deabb94" [ 771.492656] env[62066]: _type = "HttpNfcLease" [ 771.492656] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 771.498863] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 771.498863] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b52aed-9a56-02d0-a365-28d50deabb94" [ 771.498863] env[62066]: _type = "HttpNfcLease" [ 771.498863] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 771.530434] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.530434] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.530434] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 771.800180] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.800729] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.803715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 20.263s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.851597] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1155983, 'name': CloneVM_Task, 'duration_secs': 1.329336} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.851876] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created linked-clone VM from snapshot [ 771.852705] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2f17e0-c58f-4921-9a6b-8446af7ab7ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.860831] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploading image 63c467a0-7d85-4619-a71e-fe2e06314300 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 771.882874] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 771.882874] env[62066]: value = "vm-251661" [ 771.882874] env[62066]: _type = "VirtualMachine" [ 771.882874] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 771.883263] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b2577ad4-51e0-4c71-bd2e-7e8f3639eb2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.891410] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease: (returnval){ [ 771.891410] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dafe9a-a7db-8189-92b5-da55753994f2" [ 771.891410] env[62066]: _type = "HttpNfcLease" [ 771.891410] env[62066]: } obtained for exporting VM: (result){ [ 771.891410] env[62066]: value = "vm-251661" [ 771.891410] env[62066]: _type = "VirtualMachine" [ 771.891410] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 771.891823] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the lease: (returnval){ [ 771.891823] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dafe9a-a7db-8189-92b5-da55753994f2" [ 771.891823] env[62066]: _type = "HttpNfcLease" [ 771.891823] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 771.898664] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 771.898664] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dafe9a-a7db-8189-92b5-da55753994f2" [ 771.898664] env[62066]: _type = "HttpNfcLease" [ 771.898664] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 772.002017] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.002017] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b52aed-9a56-02d0-a365-28d50deabb94" [ 772.002017] env[62066]: _type = "HttpNfcLease" [ 772.002017] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 772.002325] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 772.002325] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b52aed-9a56-02d0-a365-28d50deabb94" [ 772.002325] env[62066]: _type = "HttpNfcLease" [ 772.002325] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 772.003299] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67480e79-8015-486f-b053-0726f8922ce9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.013549] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 772.014639] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 772.109698] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-93548cb6-2cd0-42f6-978c-004f6df8cbc3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.131209] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.308413] env[62066]: DEBUG nova.compute.utils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 772.313019] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 772.313019] env[62066]: DEBUG nova.objects.instance [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lazy-loading 'migration_context' on Instance uuid c0ac4362-766f-48ba-aeb2-7fd976c1f47f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 772.338319] env[62066]: DEBUG nova.network.neutron [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Updating instance_info_cache with network_info: [{"id": "88572837-2f5a-41c7-9c89-df220eebdc16", "address": "fa:16:3e:76:46:2b", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88572837-2f", "ovs_interfaceid": "88572837-2f5a-41c7-9c89-df220eebdc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.401542] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.401542] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dafe9a-a7db-8189-92b5-da55753994f2" [ 772.401542] env[62066]: _type = "HttpNfcLease" [ 772.401542] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 772.401963] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 772.401963] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dafe9a-a7db-8189-92b5-da55753994f2" [ 772.401963] env[62066]: _type = "HttpNfcLease" [ 772.401963] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 772.402905] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cea5cf-6993-41c6-888e-3ccb1010f345 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.412045] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 772.412260] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 772.531896] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fb313477-5d9f-466b-94bc-5ec885a069e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.820208] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.841678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.841678] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Instance network_info: |[{"id": "88572837-2f5a-41c7-9c89-df220eebdc16", "address": "fa:16:3e:76:46:2b", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88572837-2f", "ovs_interfaceid": "88572837-2f5a-41c7-9c89-df220eebdc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 772.844955] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:46:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6e940e5-e083-4238-973e-f1b4e2a3a5c7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88572837-2f5a-41c7-9c89-df220eebdc16', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.853919] env[62066]: DEBUG oslo.service.loopingcall [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.855104] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.855104] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d2142e0-a4d2-450b-9632-6100fb886154 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.882663] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.882663] env[62066]: value = "task-1155986" [ 772.882663] env[62066]: _type = "Task" [ 772.882663] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.896512] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155986, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.964583] env[62066]: DEBUG nova.compute.manager [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Received event network-changed-88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.965077] env[62066]: DEBUG nova.compute.manager [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Refreshing instance network info cache due to event network-changed-88572837-2f5a-41c7-9c89-df220eebdc16. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.966046] env[62066]: DEBUG oslo_concurrency.lockutils [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] Acquiring lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.966046] env[62066]: DEBUG oslo_concurrency.lockutils [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] Acquired lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.966046] env[62066]: DEBUG nova.network.neutron [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Refreshing network info cache for port 88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.291307] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7037e88f-3f0a-4a81-b1d2-a7823d2780ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.299291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77235137-a851-4c0d-9ff9-38b0ac0fff04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.337349] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01dfc75d-2885-4918-bfb4-2770f98d9605 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.345917] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37d96b3-cb99-4306-b900-b0da974652d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.362825] env[62066]: DEBUG nova.compute.provider_tree [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.393121] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155986, 'name': CreateVM_Task, 'duration_secs': 0.383552} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.393373] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.394163] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.394351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.394709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.395033] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd950392-35b7-43bd-93cc-764b2ea011a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.401148] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 773.401148] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5226c915-5c25-36a2-d502-68c98049a29b" [ 773.401148] env[62066]: _type = "Task" [ 773.401148] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.410266] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5226c915-5c25-36a2-d502-68c98049a29b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.727497] env[62066]: DEBUG nova.network.neutron [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Updated VIF entry in instance network info cache for port 88572837-2f5a-41c7-9c89-df220eebdc16. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 773.728066] env[62066]: DEBUG nova.network.neutron [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Updating instance_info_cache with network_info: [{"id": "88572837-2f5a-41c7-9c89-df220eebdc16", "address": "fa:16:3e:76:46:2b", "network": {"id": "6faab846-0412-42d6-8d1a-52c34d98f045", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-147580756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e781d81f23142d4859c188db11c59b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88572837-2f", "ovs_interfaceid": "88572837-2f5a-41c7-9c89-df220eebdc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.842247] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.866614] env[62066]: DEBUG nova.scheduler.client.report [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.912422] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5226c915-5c25-36a2-d502-68c98049a29b, 'name': SearchDatastore_Task, 'duration_secs': 0.012207} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.912732] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.912969] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.913253] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.913426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.913616] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.913876] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1256d95f-7fa7-4880-a076-fe1d68967c7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.922194] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.922398] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 773.923151] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7222a83-334d-4551-b062-80838e22c85e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.928810] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 773.928810] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]520c6d89-9717-1677-5a6f-8c381d0ab6d0" [ 773.928810] env[62066]: _type = "Task" [ 773.928810] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.936744] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]520c6d89-9717-1677-5a6f-8c381d0ab6d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.231964] env[62066]: DEBUG oslo_concurrency.lockutils [req-16eccee3-7fa2-49f7-ac8e-f4f2509dab01 req-4a72fec3-27f2-405a-9f26-fdb3b03e7918 service nova] Releasing lock "refresh_cache-c6fad95f-9167-4ad3-9c1c-16425fa1d36c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.440712] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]520c6d89-9717-1677-5a6f-8c381d0ab6d0, 'name': SearchDatastore_Task, 'duration_secs': 0.008492} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.441531] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c993df1-2981-47f9-ab5d-8441dcaaa687 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.447330] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 774.447330] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5222b269-ca1e-807d-eaa9-3ca7da894949" [ 774.447330] env[62066]: _type = "Task" [ 774.447330] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.455447] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5222b269-ca1e-807d-eaa9-3ca7da894949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.878647] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.075s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.885230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.858s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.887028] env[62066]: INFO nova.compute.claims [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.960146] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5222b269-ca1e-807d-eaa9-3ca7da894949, 'name': SearchDatastore_Task, 'duration_secs': 0.009506} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.960535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.960906] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] c6fad95f-9167-4ad3-9c1c-16425fa1d36c/c6fad95f-9167-4ad3-9c1c-16425fa1d36c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 774.961268] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7be7415-40a2-4a7b-9d4e-104dd48d2c1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.969088] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 774.969088] env[62066]: value = "task-1155987" [ 774.969088] env[62066]: _type = "Task" [ 774.969088] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.977668] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.478721] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155987, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.978963] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513044} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.979274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] c6fad95f-9167-4ad3-9c1c-16425fa1d36c/c6fad95f-9167-4ad3-9c1c-16425fa1d36c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.979490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.979742] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-852d996c-ea9f-4266-8645-6f43b6f51204 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.986837] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 775.986837] env[62066]: value = "task-1155988" [ 775.986837] env[62066]: _type = "Task" [ 775.986837] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.995360] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.370565] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392ad312-c763-45d7-8c1f-6c8caac90fa6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.378745] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c84fa1-8ea7-437b-8140-afc4a5ae7889 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.415769] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5e44d4-3711-4fd6-8c63-58da47a4fdfa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.424241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9ff255-6a89-4fde-a6b9-9622e069d731 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.429676] env[62066]: INFO nova.compute.manager [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Swapping old allocation on dict_keys(['cd4c0e36-9c88-4f73-a93c-1ff383ed97c4']) held by migration de64f433-6637-4e88-8678-c58a6683456c for instance [ 776.442374] env[62066]: DEBUG nova.compute.provider_tree [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.456251] env[62066]: DEBUG nova.scheduler.client.report [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Overwriting current allocation {'allocations': {'cd4c0e36-9c88-4f73-a93c-1ff383ed97c4': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 79}}, 'project_id': '7dba0273591846a0a9522614b16bce62', 'user_id': 'dce5e9b642424ef5b202eb2295a3fe1d', 'consumer_generation': 1} on consumer c0ac4362-766f-48ba-aeb2-7fd976c1f47f {{(pid=62066) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 776.496376] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08717} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.496648] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.497454] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d636674-47e9-4398-a4dc-ba116bc6dfea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.520421] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] c6fad95f-9167-4ad3-9c1c-16425fa1d36c/c6fad95f-9167-4ad3-9c1c-16425fa1d36c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.524226] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3728b381-ee6c-4969-ae2f-8dab9ffd4e2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.544834] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 776.544834] env[62066]: value = "task-1155989" [ 776.544834] env[62066]: _type = "Task" [ 776.544834] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.552705] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.591131] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.591343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquired lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.591524] env[62066]: DEBUG nova.network.neutron [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.948315] env[62066]: DEBUG nova.scheduler.client.report [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.054018] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155989, 'name': ReconfigVM_Task, 'duration_secs': 0.488798} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.054336] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Reconfigured VM instance instance-0000003c to attach disk [datastore2] c6fad95f-9167-4ad3-9c1c-16425fa1d36c/c6fad95f-9167-4ad3-9c1c-16425fa1d36c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.054955] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c39b3a0e-0d5e-4457-98af-fae071effb21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.062083] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 777.062083] env[62066]: value = "task-1155990" [ 777.062083] env[62066]: _type = "Task" [ 777.062083] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.062083] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.097905] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155990, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.185526] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.185856] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.190277] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.192119] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.192315] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.194478] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.194478] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.194478] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.194478] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.194478] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.196810] env[62066]: DEBUG nova.virt.hardware [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.206048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79af676b-c3ba-479f-a218-27e8bd4c3d82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.226843] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 777.233087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba681013-71c8-4194-8838-4857783834d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.240744] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d81421-c355-40e9-a8fc-2133e6baf90b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.257399] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 777.257582] env[62066]: ERROR oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk due to incomplete transfer. [ 777.258559] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c9bd4684-26d9-46c3-ad47-cd594386afb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.269362] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.275015] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Creating folder: Project (cf7ec55fe938499e85d2a8f704865fc5). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.278499] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b7fd6e7-3035-4d0e-a124-3e2eaa3f5b5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.280084] env[62066]: DEBUG oslo_vmware.rw_handles [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525028ea-61ae-660d-edb8-30ca3b644b4b/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 777.280279] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Uploaded image de8b415d-4ae3-4740-a106-cbd973431622 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 777.282508] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 777.283090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5717c199-7099-4caa-9742-c9b2c2b184c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.290143] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 777.290143] env[62066]: value = "task-1155992" [ 777.290143] env[62066]: _type = "Task" [ 777.290143] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.294222] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Created folder: Project (cf7ec55fe938499e85d2a8f704865fc5) in parent group-v251573. [ 777.294418] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Creating folder: Instances. Parent ref: group-v251663. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.294947] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0e0bab3-ac62-494a-aa50-e8c7af03dc95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.300079] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155992, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.304078] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Created folder: Instances in parent group-v251663. [ 777.304175] env[62066]: DEBUG oslo.service.loopingcall [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.304376] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.304615] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc132b1c-5968-4887-a70d-f98621692f3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.322798] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.322798] env[62066]: value = "task-1155994" [ 777.322798] env[62066]: _type = "Task" [ 777.322798] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.330630] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155994, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.454467] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.455030] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.457947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.136s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.458229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.460362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.683s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.461909] env[62066]: INFO nova.compute.claims [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.504324] env[62066]: INFO nova.scheduler.client.report [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocations for instance 6a498481-b2ca-4813-87b7-2f09dfa107f4 [ 777.561768] env[62066]: DEBUG nova.network.neutron [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [{"id": "4627c549-cbf1-40f9-94bb-0f08255025ae", "address": "fa:16:3e:96:3d:36", "network": {"id": "1426c3b5-3b85-4a72-b3db-2f446d49d935", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a95f0d02689045adbd4d942d7a467dd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d31a554-a94c-4471-892f-f65aa87b8279", "external-id": "nsx-vlan-transportzone-241", "segmentation_id": 241, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4627c549-cb", "ovs_interfaceid": "4627c549-cbf1-40f9-94bb-0f08255025ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.576095] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155990, 'name': Rename_Task, 'duration_secs': 0.173273} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.577280] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.577611] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b4bb51e-9fd7-4840-afa3-b65aa5bcb8f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.585955] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 777.585955] env[62066]: value = "task-1155995" [ 777.585955] env[62066]: _type = "Task" [ 777.585955] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.596802] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155995, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.802492] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155992, 'name': Destroy_Task, 'duration_secs': 0.503865} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.803015] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Destroyed the VM [ 777.803435] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 777.803830] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-64f843c1-d056-4c30-a4d2-8fa5cc286ae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.810466] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 777.810466] env[62066]: value = "task-1155996" [ 777.810466] env[62066]: _type = "Task" [ 777.810466] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.820970] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155996, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.835384] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1155994, 'name': CreateVM_Task, 'duration_secs': 0.34482} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.835852] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 777.836512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.836843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.838772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.839977] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4bd55b9-1997-4b63-b7d8-28f4f2e59c06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.845967] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 777.845967] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f206d1-f553-6bf2-0ed2-98154a260211" [ 777.845967] env[62066]: _type = "Task" [ 777.845967] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.857562] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f206d1-f553-6bf2-0ed2-98154a260211, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.970161] env[62066]: DEBUG nova.compute.utils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 777.974817] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.975097] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.012875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8bcc5515-ed3c-4770-95a4-bd5fdb26e4bd tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "6a498481-b2ca-4813-87b7-2f09dfa107f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.767s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.065667] env[62066]: DEBUG nova.policy [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8bcc97a71b4052b1ed4ba1b4194ff6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd61d04b8fe994dcfaa361bb87eb587d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.067888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Releasing lock "refresh_cache-c0ac4362-766f-48ba-aeb2-7fd976c1f47f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.068463] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 778.069138] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e100e032-8d02-43cf-ac02-3379df8771ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.075696] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 778.075696] env[62066]: value = "task-1155997" [ 778.075696] env[62066]: _type = "Task" [ 778.075696] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.085752] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.103929] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155995, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.321384] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155996, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.358067] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f206d1-f553-6bf2-0ed2-98154a260211, 'name': SearchDatastore_Task, 'duration_secs': 0.010967} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.358467] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.358763] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.359100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.359281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.359482] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.359820] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-038b54b0-f9c1-43c6-abf7-c3fc49f21b82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.368182] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.369135] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.369258] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1e498e4-9635-4be8-ac20-d553eb2d2060 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.377915] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 778.377915] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523b4fe5-a42e-199d-602b-b3c3829e9a55" [ 778.377915] env[62066]: _type = "Task" [ 778.377915] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.386099] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523b4fe5-a42e-199d-602b-b3c3829e9a55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.397983] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Successfully created port: c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.484308] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.589797] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155997, 'name': PowerOffVM_Task, 'duration_secs': 0.191017} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.596855] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 778.597725] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:53:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f58b5fca-fe02-4466-adf4-baeb8793402a',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-738942274',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.598025] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.598253] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.598497] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.598704] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.598903] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.599166] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.599373] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.599585] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.599780] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.600154] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.606320] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a2f5f75-8708-40df-bc3d-6d27660f5ded {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.625315] env[62066]: DEBUG oslo_vmware.api [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1155995, 'name': PowerOnVM_Task, 'duration_secs': 0.834463} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.626838] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.627277] env[62066]: INFO nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Took 8.43 seconds to spawn the instance on the hypervisor. [ 778.627612] env[62066]: DEBUG nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.628043] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 778.628043] env[62066]: value = "task-1155998" [ 778.628043] env[62066]: _type = "Task" [ 778.628043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.628913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0d6fb5-e2af-4908-bce1-651f4f7460af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.649045] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155998, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.825796] env[62066]: DEBUG oslo_vmware.api [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1155996, 'name': RemoveSnapshot_Task, 'duration_secs': 0.568757} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.830043] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 778.830762] env[62066]: INFO nova.compute.manager [None req-edd493aa-927c-4b73-bc14-1b4b06dbf401 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 15.48 seconds to snapshot the instance on the hypervisor. [ 778.896042] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523b4fe5-a42e-199d-602b-b3c3829e9a55, 'name': SearchDatastore_Task, 'duration_secs': 0.00967} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.901363] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbaf9fdb-d003-4d33-981b-fa57c23477c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.908964] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 778.908964] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527fe8a0-f14c-cd13-9517-7243437c04d3" [ 778.908964] env[62066]: _type = "Task" [ 778.908964] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.921541] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527fe8a0-f14c-cd13-9517-7243437c04d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.980645] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086de670-7d16-4a02-a7ac-11912072f455 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.991962] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff197d5f-0055-4daa-b178-71c45ef2af2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.047780] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8cd9a1-779b-4c43-995c-73a5ba52a78a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.057208] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8804e26e-29ea-4293-a36e-58f73077786a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.073662] env[62066]: DEBUG nova.compute.provider_tree [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.145318] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1155998, 'name': ReconfigVM_Task, 'duration_secs': 0.191228} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.146545] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571230e5-493d-4cfc-b855-4c2047d7c564 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.177804] env[62066]: INFO nova.compute.manager [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Took 33.90 seconds to build instance. [ 779.180629] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:53:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='f58b5fca-fe02-4466-adf4-baeb8793402a',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-738942274',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.180996] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.181257] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.181619] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.181914] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.182189] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.182534] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.182911] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.183118] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.183319] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.183501] env[62066]: DEBUG nova.virt.hardware [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.184654] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5bf32c1-d530-4b2b-bfb0-46b6eb2aa721 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.191061] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 779.191061] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c998fc-397a-3bc9-7aa1-4f2a2660fea0" [ 779.191061] env[62066]: _type = "Task" [ 779.191061] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.200434] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c998fc-397a-3bc9-7aa1-4f2a2660fea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.421650] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527fe8a0-f14c-cd13-9517-7243437c04d3, 'name': SearchDatastore_Task, 'duration_secs': 0.014592} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.421891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.422186] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.422461] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f776d33-3ada-4254-a3fc-b81d70d433cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.429429] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 779.429429] env[62066]: value = "task-1155999" [ 779.429429] env[62066]: _type = "Task" [ 779.429429] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.438108] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1155999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.500554] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.576473] env[62066]: DEBUG nova.scheduler.client.report [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.681021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-050c57c6-aa93-4c07-b1c1-3e5fb3a58dec tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.452s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.703438] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c998fc-397a-3bc9-7aa1-4f2a2660fea0, 'name': SearchDatastore_Task, 'duration_secs': 0.008864} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.710582] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfiguring VM instance instance-00000028 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 779.711035] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2de8ae0f-d089-40e7-aa7c-faceb33af0d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.732532] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 779.732532] env[62066]: value = "task-1156000" [ 779.732532] env[62066]: _type = "Task" [ 779.732532] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.742144] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156000, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.939826] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1155999, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.954332] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Successfully updated port: c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.082147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.082798] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.085889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.990s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.088273] env[62066]: INFO nova.compute.claims [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.243447] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156000, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.250713] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.250713] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.250713] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.250891] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.251825] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.251825] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.251825] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.251825] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.251998] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.252211] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.252403] env[62066]: DEBUG nova.virt.hardware [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.254352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d18361-76e1-47d2-848e-aa015f5956dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.263597] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 780.265948] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa9de4c-46b0-48dc-a899-8858a8221741 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.270425] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4da917-64ee-491e-aa83-17788bf4408a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.287416] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 780.287592] env[62066]: ERROR oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk due to incomplete transfer. [ 780.287987] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5057999a-1fed-4a92-af70-ebdd9924804a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.296328] env[62066]: DEBUG oslo_vmware.rw_handles [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bbc9c7-af1c-31e9-65cd-b80faa2a2760/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 780.296528] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Uploaded image 74f0bdd6-1197-4371-a231-52e7fe03597c to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 780.298232] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 780.298638] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d4e1acd3-2ab2-4b7b-8fa3-8ae775d65aa9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.306318] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 780.306318] env[62066]: value = "task-1156001" [ 780.306318] env[62066]: _type = "Task" [ 780.306318] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.314031] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156001, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.443088] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1155999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550249} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.443389] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.443636] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.443866] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2993c375-eb99-4fe1-a4db-7e6edc852ee9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.450877] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 780.450877] env[62066]: value = "task-1156002" [ 780.450877] env[62066]: _type = "Task" [ 780.450877] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.458972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.459123] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.459282] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.460446] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.596705] env[62066]: DEBUG nova.compute.utils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 780.597633] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 780.743798] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156000, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.788567] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.788938] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.789398] env[62066]: DEBUG nova.objects.instance [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.816462] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156001, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.963928] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147989} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.965028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.965508] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae79ae2-b610-42ee-9e4e-d94934dd088d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.988077] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.988429] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19d99851-c42f-4de4-9d1c-feb7ac41bf67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.003187] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.011662] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 781.011662] env[62066]: value = "task-1156003" [ 781.011662] env[62066]: _type = "Task" [ 781.011662] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.019766] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.102271] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.192924] env[62066]: DEBUG nova.network.neutron [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.208963] env[62066]: DEBUG nova.compute.manager [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.209823] env[62066]: DEBUG oslo_concurrency.lockutils [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.210120] env[62066]: DEBUG oslo_concurrency.lockutils [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.210357] env[62066]: DEBUG oslo_concurrency.lockutils [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.210570] env[62066]: DEBUG nova.compute.manager [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] No waiting events found dispatching network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.211154] env[62066]: WARNING nova.compute.manager [req-2c04f9d7-b4b9-4977-93b7-b31206655ff5 req-d252335f-3e03-41a6-b097-63ea2758bd2d service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received unexpected event network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef for instance with vm_state building and task_state spawning. [ 781.244710] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156000, 'name': ReconfigVM_Task, 'duration_secs': 1.425894} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.244710] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfigured VM instance instance-00000028 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 781.245867] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3768b25-8fe8-4cef-9c3e-a7fd106f7606 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.273300] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.276340] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aab966b8-f727-4c06-879f-f7e9ba82d78c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.301255] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 781.301255] env[62066]: value = "task-1156004" [ 781.301255] env[62066]: _type = "Task" [ 781.301255] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.314225] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.319990] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156001, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.379559] env[62066]: DEBUG nova.objects.instance [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.524212] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156003, 'name': ReconfigVM_Task, 'duration_secs': 0.311828} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.524518] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Reconfigured VM instance instance-0000003d to attach disk [datastore2] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.525161] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf6563a3-cd8a-4219-825a-8f641e465f97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.529750] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11da88ec-6221-4302-a9c3-e6ff637f92bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.533227] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 781.533227] env[62066]: value = "task-1156005" [ 781.533227] env[62066]: _type = "Task" [ 781.533227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.539174] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc85efbc-d197-4d97-9bbe-78c1bad8a6db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.545865] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156005, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.573233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259fbbf1-11b6-4eff-bc9b-e9c9546f224f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.580831] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd038bf3-adb3-4d0e-8d3d-4f6765c4b73a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.597237] env[62066]: DEBUG nova.compute.provider_tree [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.695592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.695953] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance network_info: |[{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 781.696476] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:3e:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e02e98f-44ce-42b7-a3ac-4034fae5d127', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9795c0a-c035-447a-9433-6cd7ff5ab2ef', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.704138] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating folder: Project (d61d04b8fe994dcfaa361bb87eb587d6). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.704371] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f142564-0de2-46d0-b49d-77b892b85edd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.715469] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created folder: Project (d61d04b8fe994dcfaa361bb87eb587d6) in parent group-v251573. [ 781.716132] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating folder: Instances. Parent ref: group-v251666. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.716132] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbac15ad-819e-4a75-aa96-48071ba64bb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.725406] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created folder: Instances in parent group-v251666. [ 781.725938] env[62066]: DEBUG oslo.service.loopingcall [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.725938] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.726048] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f57865b9-f605-495c-b5ff-645b39f2a885 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.744495] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.744495] env[62066]: value = "task-1156008" [ 781.744495] env[62066]: _type = "Task" [ 781.744495] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.752462] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156008, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.812081] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.819453] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156001, 'name': Destroy_Task, 'duration_secs': 1.410319} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.819721] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Destroyed the VM [ 781.820074] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 781.820354] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-36923e97-ebc6-494a-96d8-926fff4c2a5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.826355] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 781.826355] env[62066]: value = "task-1156009" [ 781.826355] env[62066]: _type = "Task" [ 781.826355] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.834592] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156009, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.882124] env[62066]: DEBUG nova.objects.base [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance<26d87a85-0aa3-49b9-97ca-1b7fedbebb14> lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 781.882479] env[62066]: DEBUG nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 781.927724] env[62066]: DEBUG nova.policy [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.043467] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156005, 'name': Rename_Task, 'duration_secs': 0.143154} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.043820] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.044097] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64a6aa32-4958-4c63-99e3-e952da92b822 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.052724] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 782.052724] env[62066]: value = "task-1156010" [ 782.052724] env[62066]: _type = "Task" [ 782.052724] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.065511] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.100818] env[62066]: DEBUG nova.scheduler.client.report [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.115599] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.142575] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.142821] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.142981] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.143353] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.143417] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.143548] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.143970] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.144065] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.144237] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.144455] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.144766] env[62066]: DEBUG nova.virt.hardware [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.146207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7370a4a-f9fd-4d34-8cd8-80bc1e9a51dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.158956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73de3103-8815-4520-ac73-a940e733ccb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.180958] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.186783] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Creating folder: Project (d5519c3fd7514d94981133e0fbfd0080). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.187580] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-543bfa68-f2e9-4e2d-aad4-0744fece0b50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.198892] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Created folder: Project (d5519c3fd7514d94981133e0fbfd0080) in parent group-v251573. [ 782.199146] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Creating folder: Instances. Parent ref: group-v251669. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.199441] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29970f48-1c6b-4af4-8687-55b0a808fca8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.210542] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Created folder: Instances in parent group-v251669. [ 782.210651] env[62066]: DEBUG oslo.service.loopingcall [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.210986] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.211195] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de6acb1a-09f1-450f-8267-d984b1497b3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.230303] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.230303] env[62066]: value = "task-1156013" [ 782.230303] env[62066]: _type = "Task" [ 782.230303] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.242584] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156013, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.254675] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156008, 'name': CreateVM_Task, 'duration_secs': 0.344412} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.254947] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.255658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.255837] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.256510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.256875] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65f7c93b-a987-45d8-99e3-f3e6b423b761 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.262043] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 782.262043] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ee83-af63-9a3a-59df-77823a4eec65" [ 782.262043] env[62066]: _type = "Task" [ 782.262043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.271651] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ee83-af63-9a3a-59df-77823a4eec65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.303261] env[62066]: DEBUG nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Successfully created port: 8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.315478] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156004, 'name': ReconfigVM_Task, 'duration_secs': 0.622109} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.315979] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f/c0ac4362-766f-48ba-aeb2-7fd976c1f47f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.316855] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e7ff87-6c76-4052-947a-a6eb8c56ff6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.339872] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6350ed6b-17c8-47bf-91b1-7e8d03118dcc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.342771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "92b150af-351a-43ab-952a-209225733ac8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.343039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.343252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "92b150af-351a-43ab-952a-209225733ac8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.343466] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.343648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.346572] env[62066]: INFO nova.compute.manager [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Terminating instance [ 782.365623] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156009, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.367912] env[62066]: DEBUG nova.compute.manager [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.367912] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.368744] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8639423f-ed5a-4559-9dc4-07c7088a796d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.372128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab63092e-03fd-4633-b154-74beac0e87be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.400824] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6830f9-6bdc-492a-99cf-fcda657d6c68 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.404232] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.404640] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-116e3ddf-2d51-45e6-9ad2-b11952501353 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.410867] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.412563] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b12a91ee-9b43-4bed-aa01-57c0cdcb1f4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.414826] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 782.414826] env[62066]: value = "task-1156014" [ 782.414826] env[62066]: _type = "Task" [ 782.414826] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.422717] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 782.422717] env[62066]: value = "task-1156015" [ 782.422717] env[62066]: _type = "Task" [ 782.422717] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.431990] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.438296] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156015, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.563612] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156010, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.607471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.607834] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.612159] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.983s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.614698] env[62066]: INFO nova.compute.claims [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.650949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "248f5281-eb46-4be1-8642-28813c4b2622" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.651329] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.651597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "248f5281-eb46-4be1-8642-28813c4b2622-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.651927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.652161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.659710] env[62066]: INFO nova.compute.manager [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Terminating instance [ 782.662413] env[62066]: DEBUG nova.compute.manager [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.662413] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.665067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d4e588-3543-4649-a6da-d36e7cbf62e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.674146] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.674462] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e7ce378-0be8-4fd8-b373-a593623c03f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.741472] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156013, 'name': CreateVM_Task, 'duration_secs': 0.31166} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.741472] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.741472] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.772072] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ee83-af63-9a3a-59df-77823a4eec65, 'name': SearchDatastore_Task, 'duration_secs': 0.020068} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.772401] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.772640] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.772879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.773041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.773228] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.773673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.773821] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.774062] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06a6d4c6-532b-4164-b37f-18bb4371a9f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.776142] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc038df5-bf34-4677-b8c5-256016b024b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.780933] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 782.780933] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52786ef2-6d41-f090-9cbd-eb104430d5ee" [ 782.780933] env[62066]: _type = "Task" [ 782.780933] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.784775] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.784960] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 782.785948] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb1e014b-78ce-46ab-8260-5cfb640186a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.791192] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52786ef2-6d41-f090-9cbd-eb104430d5ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.794076] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 782.794076] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5274e36a-5cd1-0cd0-59c9-741961051e76" [ 782.794076] env[62066]: _type = "Task" [ 782.794076] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.801426] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5274e36a-5cd1-0cd0-59c9-741961051e76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.846087] env[62066]: DEBUG oslo_vmware.api [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156009, 'name': RemoveSnapshot_Task, 'duration_secs': 0.902337} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.846146] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 782.846484] env[62066]: INFO nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Took 15.54 seconds to snapshot the instance on the hypervisor. [ 782.879344] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.879912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.880392] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.880650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.880892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.883962] env[62066]: INFO nova.compute.manager [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Terminating instance [ 782.886052] env[62066]: DEBUG nova.compute.manager [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.886265] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.888292] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcf9535-5138-47bf-93f4-9296ea861a16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.896276] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.896563] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c67499a1-984d-4a2d-8c00-a42b7885c85e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.903623] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 782.903623] env[62066]: value = "task-1156017" [ 782.903623] env[62066]: _type = "Task" [ 782.903623] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.912400] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.924936] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156014, 'name': PowerOffVM_Task, 'duration_secs': 0.219979} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.931029] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 782.931029] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.931312] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a60e4570-3484-4e67-b465-19d67e1c7f75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.939498] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156015, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.971230] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.971565] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.971828] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleting the datastore file [datastore1] 248f5281-eb46-4be1-8642-28813c4b2622 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.972177] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d17e0a02-0f8d-494b-9f39-1c241fd830e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.979053] env[62066]: DEBUG oslo_vmware.api [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 782.979053] env[62066]: value = "task-1156019" [ 782.979053] env[62066]: _type = "Task" [ 782.979053] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.988452] env[62066]: DEBUG oslo_vmware.api [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.996645] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.996878] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.997092] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleting the datastore file [datastore1] 92b150af-351a-43ab-952a-209225733ac8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.997371] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd60b032-0135-49e0-ad6d-94ab64e83ac9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.005215] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 783.005215] env[62066]: value = "task-1156020" [ 783.005215] env[62066]: _type = "Task" [ 783.005215] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.014670] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.064189] env[62066]: DEBUG oslo_vmware.api [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156010, 'name': PowerOnVM_Task, 'duration_secs': 0.518105} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.064543] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.064847] env[62066]: INFO nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Took 9.22 seconds to spawn the instance on the hypervisor. [ 783.065178] env[62066]: DEBUG nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.066184] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbd94dc-4b50-46b8-857f-f729d6b103fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.120288] env[62066]: DEBUG nova.compute.utils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 783.125023] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.125023] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.165893] env[62066]: DEBUG nova.policy [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.291651] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52786ef2-6d41-f090-9cbd-eb104430d5ee, 'name': SearchDatastore_Task, 'duration_secs': 0.015091} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.291944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.292230] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.292452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.304661] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5274e36a-5cd1-0cd0-59c9-741961051e76, 'name': SearchDatastore_Task, 'duration_secs': 0.019619} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.305511] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ff0e57a-9f10-45a4-9c7b-6433a5db0cd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.311055] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 783.311055] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524ef55a-aab4-81fe-2654-5f471f3e8e84" [ 783.311055] env[62066]: _type = "Task" [ 783.311055] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.319127] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524ef55a-aab4-81fe-2654-5f471f3e8e84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.351026] env[62066]: DEBUG nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance disappeared during snapshot {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 783.362161] env[62066]: DEBUG nova.compute.manager [None req-930ec3d1-2dc1-42a2-925d-cc73d6f2b1f9 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image not found during clean up 74f0bdd6-1197-4371-a231-52e7fe03597c {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 783.416033] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156017, 'name': PowerOffVM_Task, 'duration_secs': 0.237819} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.416033] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.416033] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 783.416234] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9baf045b-96c5-4040-a484-a2c3457725ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.430601] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Successfully created port: 9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.435885] env[62066]: DEBUG oslo_vmware.api [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156015, 'name': PowerOnVM_Task, 'duration_secs': 0.594781} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.436191] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.487466] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 783.487734] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 783.487941] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleting the datastore file [datastore2] c6fad95f-9167-4ad3-9c1c-16425fa1d36c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.493353] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cde45f65-c7e4-45f3-80c6-69c93e78667e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.493540] env[62066]: DEBUG oslo_vmware.api [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156019, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326198} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.493734] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 783.493994] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 783.494213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 783.494391] env[62066]: INFO nova.compute.manager [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Took 0.83 seconds to destroy the instance on the hypervisor. [ 783.494630] env[62066]: DEBUG oslo.service.loopingcall [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.495211] env[62066]: DEBUG nova.compute.manager [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.495304] env[62066]: DEBUG nova.network.neutron [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.500571] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for the task: (returnval){ [ 783.500571] env[62066]: value = "task-1156022" [ 783.500571] env[62066]: _type = "Task" [ 783.500571] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.509131] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.516472] env[62066]: DEBUG oslo_vmware.api [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302947} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.516719] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 783.516902] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 783.517103] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 783.518494] env[62066]: INFO nova.compute.manager [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: 92b150af-351a-43ab-952a-209225733ac8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 783.518494] env[62066]: DEBUG oslo.service.loopingcall [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.518494] env[62066]: DEBUG nova.compute.manager [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.518494] env[62066]: DEBUG nova.network.neutron [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.584770] env[62066]: INFO nova.compute.manager [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Took 33.81 seconds to build instance. [ 783.624814] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.828807] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524ef55a-aab4-81fe-2654-5f471f3e8e84, 'name': SearchDatastore_Task, 'duration_secs': 0.022698} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.829149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.829438] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.829796] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.830123] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.830424] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a868a1e4-e666-4a2c-a59b-40508bb0fca3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.835983] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f0c16ea-5b7c-4087-9642-d2f1bb15ea34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.842819] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 783.842819] env[62066]: value = "task-1156023" [ 783.842819] env[62066]: _type = "Task" [ 783.842819] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.848577] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.848780] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.852154] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a483a725-ed43-4233-b123-cb9f88d0f0f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.859953] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.862589] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "854a1ea4-515d-4a05-b179-22713b63f7c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.862835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.866900] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 783.866900] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52afec4c-fcae-84cb-b064-562c32e6ebcc" [ 783.866900] env[62066]: _type = "Task" [ 783.866900] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.880895] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52afec4c-fcae-84cb-b064-562c32e6ebcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.009472] env[62066]: DEBUG oslo_vmware.api [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Task: {'id': task-1156022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379414} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.009705] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.009897] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.010218] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.010980] env[62066]: INFO nova.compute.manager [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 784.010980] env[62066]: DEBUG oslo.service.loopingcall [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.012139] env[62066]: DEBUG nova.compute.manager [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.012139] env[62066]: DEBUG nova.network.neutron [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.092937] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c8b372d2-64e8-4662-89fe-50d7bf1cfd46 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.861s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.157044] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c7d791-1be7-4ffe-908b-2273d5c4362e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.166875] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47c83f5-8fd0-4413-b81c-03a773bcec7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.210493] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc4a7fc-a9ab-4d51-a16a-b8f98ebb7bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.219680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c110a09-eecb-42bd-b097-4fb4f13cace4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.238183] env[62066]: DEBUG nova.compute.provider_tree [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.355824] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156023, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.364986] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.379718] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52afec4c-fcae-84cb-b064-562c32e6ebcc, 'name': SearchDatastore_Task, 'duration_secs': 0.018023} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.380678] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eaf17ff-4422-45de-8a50-637e4186c0a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.388473] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 784.388473] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525e8636-0f60-0f1d-fc01-4f63ccd93294" [ 784.388473] env[62066]: _type = "Task" [ 784.388473] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.397867] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525e8636-0f60-0f1d-fc01-4f63ccd93294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.452537] env[62066]: INFO nova.compute.manager [None req-a8b98e0e-e234-4de2-ae64-8421a29eb056 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance to original state: 'active' [ 784.460307] env[62066]: DEBUG nova.compute.manager [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.460544] env[62066]: DEBUG nova.compute.manager [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing instance network info cache due to event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.460776] env[62066]: DEBUG oslo_concurrency.lockutils [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.460926] env[62066]: DEBUG oslo_concurrency.lockutils [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.461424] env[62066]: DEBUG nova.network.neutron [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.533019] env[62066]: DEBUG nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Successfully updated port: 8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.643240] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.665520] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.665787] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.665947] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.666182] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.666340] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.666491] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.666700] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.666861] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.667470] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.667723] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.668059] env[62066]: DEBUG nova.virt.hardware [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.668925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef6ab32-fb9e-48e5-8a08-1a244a4f4421 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.678641] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349e67df-a5e4-44c4-a811-c34a9bda23ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.742428] env[62066]: DEBUG nova.scheduler.client.report [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.817384] env[62066]: DEBUG nova.network.neutron [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.823949] env[62066]: DEBUG nova.network.neutron [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.838314] env[62066]: DEBUG nova.network.neutron [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.854478] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683134} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.854755] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.854993] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.855255] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38f74f99-113e-45df-a2f6-c3e553f0fce3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.862030] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 784.862030] env[62066]: value = "task-1156024" [ 784.862030] env[62066]: _type = "Task" [ 784.862030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.870176] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.895661] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.900818] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525e8636-0f60-0f1d-fc01-4f63ccd93294, 'name': SearchDatastore_Task, 'duration_secs': 0.053421} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.900915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.901382] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] e95ee795-603e-4cbf-bcd6-1ba54b62a281/e95ee795-603e-4cbf-bcd6-1ba54b62a281.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.901500] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-465967ef-5bac-453c-90d6-bec54ee24427 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.908648] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 784.908648] env[62066]: value = "task-1156025" [ 784.908648] env[62066]: _type = "Task" [ 784.908648] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.916986] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.937151] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Successfully updated port: 9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.036049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.036049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.036049] env[62066]: DEBUG nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.209680] env[62066]: DEBUG nova.network.neutron [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updated VIF entry in instance network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.209680] env[62066]: DEBUG nova.network.neutron [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.235939] env[62066]: DEBUG nova.compute.manager [req-0f12e1a7-0c72-41e8-bed5-fec89b3e371d req-677d432a-0fe0-4fe4-9e81-075bae6f671d service nova] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Received event network-vif-deleted-88572837-2f5a-41c7-9c89-df220eebdc16 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.251473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.251473] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.252907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.089s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.253189] env[62066]: DEBUG nova.objects.instance [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lazy-loading 'resources' on Instance uuid 641bf537-76d2-4c41-a6dc-b5b042327477 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.319665] env[62066]: INFO nova.compute.manager [-] [instance: 92b150af-351a-43ab-952a-209225733ac8] Took 1.80 seconds to deallocate network for instance. [ 785.325581] env[62066]: INFO nova.compute.manager [-] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Took 1.83 seconds to deallocate network for instance. [ 785.340798] env[62066]: INFO nova.compute.manager [-] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Took 1.33 seconds to deallocate network for instance. [ 785.375471] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065295} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.375847] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.377057] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a937cd78-6da1-4281-8b3b-8b549c592114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.409789] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.411032] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d4a5a70-51e2-417d-b750-e17c5ef15473 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.439539] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.439660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.439776] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.441401] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156025, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.443301] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 785.443301] env[62066]: value = "task-1156026" [ 785.443301] env[62066]: _type = "Task" [ 785.443301] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.454498] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156026, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.571891] env[62066]: WARNING nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 785.713045] env[62066]: DEBUG oslo_concurrency.lockutils [req-be4290f5-857a-42d5-a340-c73cd42794f7 req-60e6824a-edd0-4ef5-858a-4b579d90a383 service nova] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.761663] env[62066]: DEBUG nova.compute.utils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 785.766035] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.766035] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.814113] env[62066]: DEBUG nova.policy [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb622b6e53894cec9c0f9a809dfcd3a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '720fe50954804162a95423afa176e526', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.827727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.841496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.849797] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.923748] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592687} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.923966] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] e95ee795-603e-4cbf-bcd6-1ba54b62a281/e95ee795-603e-4cbf-bcd6-1ba54b62a281.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.924675] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.928425] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b814f1e2-139b-4a6e-8e66-60f61c0bb90a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.936774] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 785.936774] env[62066]: value = "task-1156027" [ 785.936774] env[62066]: _type = "Task" [ 785.936774] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.952364] env[62066]: DEBUG nova.network.neutron [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5014e0-3b5b-46a1-b1b6-b98e168c945f", "address": "fa:16:3e:7d:78:a0", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5014e0-3b", "ovs_interfaceid": "8a5014e0-3b5b-46a1-b1b6-b98e168c945f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.957567] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.963689] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156026, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.987290] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.028851] env[62066]: DEBUG nova.objects.instance [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lazy-loading 'flavor' on Instance uuid 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 786.163920] env[62066]: DEBUG nova.network.neutron [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Updating instance_info_cache with network_info: [{"id": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "address": "fa:16:3e:cc:1b:73", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a995991-9d", "ovs_interfaceid": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.194970] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Successfully created port: 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.249645] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19605bf9-0a23-4f59-b9b2-39e3134dbea4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.258173] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7431b9ec-c8c7-48d4-8863-daeae91c3a2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.292020] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.295526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15902352-b6c4-4fc9-9071-4e90bd7ab918 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.304185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826d7b7e-fae7-4d10-9aa4-cd97662924eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.321626] env[62066]: DEBUG nova.compute.provider_tree [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.446746] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156027, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.176707} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.449717] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.451468] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47598ac-fd81-47ee-b84c-b0ea97765b2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.461089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.461089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.461089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.461089] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156026, 'name': ReconfigVM_Task, 'duration_secs': 0.691828} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.470397] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87099723-ac66-4125-8021-49888756b08d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.472896] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.483445] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] e95ee795-603e-4cbf-bcd6-1ba54b62a281/e95ee795-603e-4cbf-bcd6-1ba54b62a281.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.483745] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-131d0e82-de28-45b6-bb7f-541e489e8a71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.485354] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18081fe0-a07d-472d-af9a-6df6ab00dd66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.514569] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.514831] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.514993] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.515196] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.515351] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.515560] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.515847] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.516044] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.516222] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.516390] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.516567] env[62066]: DEBUG nova.virt.hardware [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.523128] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 786.526691] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbf9a347-f3be-4e44-a867-e7a9391e08f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.539152] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 786.539152] env[62066]: value = "task-1156029" [ 786.539152] env[62066]: _type = "Task" [ 786.539152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.539431] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 786.539431] env[62066]: value = "task-1156028" [ 786.539431] env[62066]: _type = "Task" [ 786.539431] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.554799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.554989] env[62066]: DEBUG oslo_concurrency.lockutils [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.557484] env[62066]: DEBUG oslo_vmware.api [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 786.557484] env[62066]: value = "task-1156030" [ 786.557484] env[62066]: _type = "Task" [ 786.557484] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.568688] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156028, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.568954] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156029, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.575323] env[62066]: DEBUG oslo_vmware.api [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156030, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.667029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.667357] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Instance network_info: |[{"id": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "address": "fa:16:3e:cc:1b:73", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a995991-9d", "ovs_interfaceid": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 786.669874] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:1b:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a995991-9d98-4fcc-a69b-cd7d9209d89f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.677073] env[62066]: DEBUG oslo.service.loopingcall [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.677073] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 786.677261] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c8e5c08-c512-421d-a9a6-9259cd0cdea8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.700231] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.700231] env[62066]: value = "task-1156031" [ 786.700231] env[62066]: _type = "Task" [ 786.700231] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.711233] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156031, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.824775] env[62066]: DEBUG nova.scheduler.client.report [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.050249] env[62066]: INFO nova.compute.manager [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Rebuilding instance [ 787.073041] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156029, 'name': ReconfigVM_Task, 'duration_secs': 0.327249} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.073041] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156028, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.073041] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Reconfigured VM instance instance-0000003f to attach disk [datastore2] e95ee795-603e-4cbf-bcd6-1ba54b62a281/e95ee795-603e-4cbf-bcd6-1ba54b62a281.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.073041] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20f42328-6980-48b5-be5a-58fcfe600d00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.078375] env[62066]: DEBUG oslo_vmware.api [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.083876] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 787.083876] env[62066]: value = "task-1156032" [ 787.083876] env[62066]: _type = "Task" [ 787.083876] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.103161] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156032, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.107741] env[62066]: DEBUG nova.compute.manager [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.109297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2233144-f754-447d-a1c6-45cacc43b133 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 92b150af-351a-43ab-952a-209225733ac8] Received event network-vif-deleted-4f7a82c6-2b47-414a-9821-e48187f88ba3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Received event network-vif-deleted-58c12e67-83e3-4f98-bbc7-b90eef01f919 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-vif-plugged-8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.193264] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.193264] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.193264] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] No waiting events found dispatching network-vif-plugged-8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.193264] env[62066]: WARNING nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received unexpected event network-vif-plugged-8a5014e0-3b5b-46a1-b1b6-b98e168c945f for instance with vm_state active and task_state None. [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-changed-8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.193264] env[62066]: DEBUG nova.compute.manager [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing instance network info cache due to event network-changed-8a5014e0-3b5b-46a1-b1b6-b98e168c945f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.193264] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.193264] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.194145] env[62066]: DEBUG nova.network.neutron [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Refreshing network info cache for port 8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.211736] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156031, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.236427] env[62066]: DEBUG nova.network.neutron [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.305753] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.319305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.319613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.319846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.320090] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.320523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.322623] env[62066]: INFO nova.compute.manager [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Terminating instance [ 787.327055] env[62066]: DEBUG nova.compute.manager [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 787.327234] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.328677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0c3ee3-cf98-406c-b02f-ceb3e5b60171 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.332410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.080s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.337283] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.338052] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.338052] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.338052] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.338268] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.338450] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.338527] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.339129] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.339129] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.339129] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.339320] env[62066]: DEBUG nova.virt.hardware [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.339548] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.407s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.341157] env[62066]: INFO nova.compute.claims [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.345073] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99192e9-d298-4aa5-b913-1b806994d0ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.350904] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 787.351595] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51dd4c1c-c759-4df5-8e92-6386f1c00bcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.357269] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7525e8a3-4b49-4644-9320-c73e720d42ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.362900] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 787.362900] env[62066]: value = "task-1156033" [ 787.362900] env[62066]: _type = "Task" [ 787.362900] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.376306] env[62066]: INFO nova.scheduler.client.report [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Deleted allocations for instance 641bf537-76d2-4c41-a6dc-b5b042327477 [ 787.385041] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.534958] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 787.535955] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1679c39d-968e-412b-9547-db4ef7d5ae13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.543234] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 787.543477] env[62066]: ERROR oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk due to incomplete transfer. [ 787.543741] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-54d12514-3e4e-4638-bf83-895780b927a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.551524] env[62066]: DEBUG oslo_vmware.rw_handles [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db9921-d9ba-6988-a537-6a2286a00467/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 787.551797] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploaded image 63c467a0-7d85-4619-a71e-fe2e06314300 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 787.553566] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 787.554182] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ca47e610-e11e-400e-aa3d-0b9d6e738b74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.559406] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156028, 'name': Rename_Task, 'duration_secs': 0.802276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.562404] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.564537] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-347fd709-a7df-4d90-9d73-06a116e091fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.565894] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 787.565894] env[62066]: value = "task-1156034" [ 787.565894] env[62066]: _type = "Task" [ 787.565894] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.573368] env[62066]: DEBUG oslo_vmware.api [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.574961] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 787.574961] env[62066]: value = "task-1156035" [ 787.574961] env[62066]: _type = "Task" [ 787.574961] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.581484] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156034, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.586516] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.597698] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156032, 'name': Rename_Task, 'duration_secs': 0.155744} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.598578] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.598578] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1da06223-ef97-4669-9148-74ee88c2ca8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.604873] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 787.604873] env[62066]: value = "task-1156036" [ 787.604873] env[62066]: _type = "Task" [ 787.604873] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.612838] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.622580] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 787.622892] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a66145da-bf6c-4741-b0c8-e3c04fc6c8ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.629281] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 787.629281] env[62066]: value = "task-1156037" [ 787.629281] env[62066]: _type = "Task" [ 787.629281] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.638421] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.712860] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156031, 'name': CreateVM_Task, 'duration_secs': 0.752683} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.713148] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.714112] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.714326] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.714746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.715034] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbe2087d-cb03-4e4f-8f1a-3eda47b1a771 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.722626] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 787.722626] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527a6443-f069-896e-fc57-521517526582" [ 787.722626] env[62066]: _type = "Task" [ 787.722626] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.731500] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a6443-f069-896e-fc57-521517526582, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.849403] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Received event network-vif-plugged-9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.849651] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Acquiring lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.849986] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.850414] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.850676] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] No waiting events found dispatching network-vif-plugged-9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.851171] env[62066]: WARNING nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Received unexpected event network-vif-plugged-9a995991-9d98-4fcc-a69b-cd7d9209d89f for instance with vm_state building and task_state spawning. [ 787.855206] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Received event network-changed-9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.855438] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Refreshing instance network info cache due to event network-changed-9a995991-9d98-4fcc-a69b-cd7d9209d89f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.855656] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Acquiring lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.855801] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Acquired lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.855998] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Refreshing network info cache for port 9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.882792] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156033, 'name': PowerOffVM_Task, 'duration_secs': 0.257476} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.884607] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Successfully updated port: 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.884935] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 787.885162] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 787.891017] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-802f6fdb-9623-4187-b1ba-155e89316f17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.897203] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6189b93f-cab4-4918-bc56-9f265ad80ae1 tempest-SecurityGroupsTestJSON-797402054 tempest-SecurityGroupsTestJSON-797402054-project-member] Lock "641bf537-76d2-4c41-a6dc-b5b042327477" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.202s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.974170] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.974491] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.974749] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Deleting the datastore file [datastore2] c0ac4362-766f-48ba-aeb2-7fd976c1f47f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.975051] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-681a927f-d588-46e8-818c-3e6347d7bf39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.981909] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for the task: (returnval){ [ 787.981909] env[62066]: value = "task-1156039" [ 787.981909] env[62066]: _type = "Task" [ 787.981909] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.994086] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.069306] env[62066]: DEBUG oslo_vmware.api [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156030, 'name': ReconfigVM_Task, 'duration_secs': 1.176715} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.073892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.073892] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 788.076161] env[62066]: DEBUG nova.network.neutron [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updated VIF entry in instance network info cache for port 8a5014e0-3b5b-46a1-b1b6-b98e168c945f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.076586] env[62066]: DEBUG nova.network.neutron [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "8a5014e0-3b5b-46a1-b1b6-b98e168c945f", "address": "fa:16:3e:7d:78:a0", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a5014e0-3b", "ovs_interfaceid": "8a5014e0-3b5b-46a1-b1b6-b98e168c945f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.084777] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156034, 'name': Destroy_Task, 'duration_secs': 0.350198} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.089240] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroyed the VM [ 788.089240] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 788.089240] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-93cb6b47-30b0-4211-b389-d955b233dc25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.095028] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156035, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.099644] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 788.099644] env[62066]: value = "task-1156040" [ 788.099644] env[62066]: _type = "Task" [ 788.099644] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.112915] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156040, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.117923] env[62066]: DEBUG oslo_vmware.api [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156036, 'name': PowerOnVM_Task, 'duration_secs': 0.474081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.119073] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.119621] env[62066]: INFO nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Took 6.00 seconds to spawn the instance on the hypervisor. [ 788.119621] env[62066]: DEBUG nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.120542] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429e03eb-3db7-4252-9891-39ef7fdbaeb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.130637] env[62066]: DEBUG nova.network.neutron [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.141551] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156037, 'name': PowerOffVM_Task, 'duration_secs': 0.135301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.141833] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.142121] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.143036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d4c5ce-2bde-41ea-b32b-5f8b8795d3b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.151048] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 788.151048] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e8ffcc5-cc2c-48b3-ac3a-07406669ceb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.179021] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 788.179298] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 788.179481] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Deleting the datastore file [datastore2] d9163e64-309b-4381-8819-15757f83ac2e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.179742] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-879aa5af-b3df-4880-b8c3-0e8e9e07c274 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.185925] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 788.185925] env[62066]: value = "task-1156042" [ 788.185925] env[62066]: _type = "Task" [ 788.185925] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.196525] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.233473] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a6443-f069-896e-fc57-521517526582, 'name': SearchDatastore_Task, 'duration_secs': 0.01216} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.233813] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.234087] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.234338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.234559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.234775] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.235087] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b5ede02-f8cd-4ad2-8670-2affc758e114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.243203] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.244170] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.244325] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a35b71-e83b-45a8-96bd-535af9279a85 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.250917] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 788.250917] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]520ccee4-97b1-3544-1a08-fe0c523f977d" [ 788.250917] env[62066]: _type = "Task" [ 788.250917] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.259712] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]520ccee4-97b1-3544-1a08-fe0c523f977d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.394819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.394819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.394819] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.495629] env[62066]: DEBUG oslo_vmware.api [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Task: {'id': task-1156039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.290154} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.495897] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 788.496406] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 788.496406] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 788.496502] env[62066]: INFO nova.compute.manager [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 788.496749] env[62066]: DEBUG oslo.service.loopingcall [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.496908] env[62066]: DEBUG nova.compute.manager [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.497404] env[62066]: DEBUG nova.network.neutron [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 788.588148] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c63a2ce0-baff-4ae8-b5f6-4119369f0c98 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.799s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.589376] env[62066]: DEBUG oslo_concurrency.lockutils [req-1075c574-4818-4447-a81f-69969ee40784 req-f9c3e438-be51-4bce-b3b0-cc169999cacd service nova] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.589799] env[62066]: DEBUG oslo_vmware.api [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156035, 'name': PowerOnVM_Task, 'duration_secs': 0.684371} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.594237] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.594447] env[62066]: INFO nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Took 9.09 seconds to spawn the instance on the hypervisor. [ 788.594629] env[62066]: DEBUG nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.597650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce652550-2337-462b-b480-02145405ebe8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.621384] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156040, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.637541] env[62066]: DEBUG oslo_concurrency.lockutils [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.637856] env[62066]: DEBUG nova.compute.manager [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Inject network info {{(pid=62066) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 788.638106] env[62066]: DEBUG nova.compute.manager [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] network_info to inject: |[{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 788.643171] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfiguring VM instance to set the machine id {{(pid=62066) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 788.648226] env[62066]: INFO nova.compute.manager [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Took 32.89 seconds to build instance. [ 788.650339] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-863929b9-b47b-4029-8375-5fae81e2f6dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.672421] env[62066]: DEBUG oslo_vmware.api [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 788.672421] env[62066]: value = "task-1156043" [ 788.672421] env[62066]: _type = "Task" [ 788.672421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.678664] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Updated VIF entry in instance network info cache for port 9a995991-9d98-4fcc-a69b-cd7d9209d89f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.679433] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Updating instance_info_cache with network_info: [{"id": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "address": "fa:16:3e:cc:1b:73", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a995991-9d", "ovs_interfaceid": "9a995991-9d98-4fcc-a69b-cd7d9209d89f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.687643] env[62066]: DEBUG oslo_vmware.api [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156043, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.698323] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16962} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.699500] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 788.699500] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 788.699500] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 788.761816] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]520ccee4-97b1-3544-1a08-fe0c523f977d, 'name': SearchDatastore_Task, 'duration_secs': 0.019696} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.765289] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-614fb362-d742-4aec-a92e-57993cbc67c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.771649] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 788.771649] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521a303d-b759-ea3b-213d-d4ebf4161f0f" [ 788.771649] env[62066]: _type = "Task" [ 788.771649] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.785152] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521a303d-b759-ea3b-213d-d4ebf4161f0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.830928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6dbe83-9dc9-48e0-9e50-d372095e3fef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.842642] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d45db69-2134-4b2d-8cc1-658a4b217506 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.901265] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da86cc97-a120-4e8c-8ec6-406f3296dd62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.914076] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbc1c5b-78c6-4c8e-8fad-6d0e19f455d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.937140] env[62066]: DEBUG nova.compute.provider_tree [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.939668] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.065436] env[62066]: DEBUG nova.network.neutron [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.111404] env[62066]: DEBUG oslo_vmware.api [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156040, 'name': RemoveSnapshot_Task, 'duration_secs': 0.799403} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.111404] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 789.111404] env[62066]: INFO nova.compute.manager [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 20.88 seconds to snapshot the instance on the hypervisor. [ 789.132216] env[62066]: INFO nova.compute.manager [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Took 36.12 seconds to build instance. [ 789.152573] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b8e3d59-455e-48b5-95e8-35ad09632f8b tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.862s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.182715] env[62066]: DEBUG oslo_vmware.api [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156043, 'name': ReconfigVM_Task, 'duration_secs': 0.151699} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.182984] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-92973991-4179-43a0-a658-f5c63ae34fc0 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfigured VM instance to set the machine id {{(pid=62066) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 789.189037] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Releasing lock "refresh_cache-43d64f76-7f7a-4b95-b9df-c95218612998" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.189283] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.189460] env[62066]: DEBUG nova.compute.manager [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing instance network info cache due to event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.189674] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.189821] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.189986] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.284096] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521a303d-b759-ea3b-213d-d4ebf4161f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.014298} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.284408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.284675] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 43d64f76-7f7a-4b95-b9df-c95218612998/43d64f76-7f7a-4b95-b9df-c95218612998.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.284934] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6b4fa89-fc2d-4fd6-b03d-0f416217a9fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.294020] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 789.294020] env[62066]: value = "task-1156044" [ 789.294020] env[62066]: _type = "Task" [ 789.294020] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.302312] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.336882] env[62066]: DEBUG nova.network.neutron [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.447027] env[62066]: DEBUG nova.scheduler.client.report [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.550639] env[62066]: DEBUG nova.objects.instance [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lazy-loading 'flavor' on Instance uuid 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.570015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.570015] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Instance network_info: |[{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.570015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:1e:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d1d88f2-e881-41ed-8de8-7cf13954abf9', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.576815] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Creating folder: Project (720fe50954804162a95423afa176e526). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 789.577447] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62598529-201e-4db8-9aec-2e74facc24e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.588623] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Created folder: Project (720fe50954804162a95423afa176e526) in parent group-v251573. [ 789.589226] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Creating folder: Instances. Parent ref: group-v251673. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 789.589601] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2447f0c4-7d72-4bc2-b783-99f8b2011b7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.599043] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Created folder: Instances in parent group-v251673. [ 789.599043] env[62066]: DEBUG oslo.service.loopingcall [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.599043] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 789.599155] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5bc985f-4d5f-4fe9-b403-dc72386a250b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.621636] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.621636] env[62066]: value = "task-1156047" [ 789.621636] env[62066]: _type = "Task" [ 789.621636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.630158] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156047, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.633788] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3822718f-b096-4ce6-aee2-6ede9722c34c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.393s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.668134] env[62066]: DEBUG nova.compute.manager [None req-228274ef-f0ea-44d3-88ad-8f35660f658e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Found 1 images (rotation: 2) {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 789.736052] env[62066]: DEBUG nova.compute.manager [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-vif-plugged-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.736300] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Acquiring lock "4723be94-3479-4e66-8088-914824c0e669-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.736522] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Lock "4723be94-3479-4e66-8088-914824c0e669-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.736695] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Lock "4723be94-3479-4e66-8088-914824c0e669-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.736863] env[62066]: DEBUG nova.compute.manager [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] No waiting events found dispatching network-vif-plugged-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 789.737047] env[62066]: WARNING nova.compute.manager [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received unexpected event network-vif-plugged-4d1d88f2-e881-41ed-8de8-7cf13954abf9 for instance with vm_state building and task_state spawning. [ 789.737223] env[62066]: DEBUG nova.compute.manager [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.737384] env[62066]: DEBUG nova.compute.manager [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing instance network info cache due to event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.737573] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.737713] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.737873] env[62066]: DEBUG nova.network.neutron [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.742089] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.742197] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.742410] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.742746] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.742936] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.743118] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.743342] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.743511] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.743683] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.743855] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.744043] env[62066]: DEBUG nova.virt.hardware [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.748102] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d8448a-ba26-4b37-81d8-9e0d8bc5264f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.755520] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8307df-64a7-4b9e-b6d7-9fb149020286 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.772305] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.778419] env[62066]: DEBUG oslo.service.loopingcall [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.779255] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 789.779547] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0e712de-6fc5-4a55-9752-84e1d555c38c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.803207] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.803207] env[62066]: value = "task-1156048" [ 789.803207] env[62066]: _type = "Task" [ 789.803207] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.813562] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156044, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.820010] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156048, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.840342] env[62066]: INFO nova.compute.manager [-] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Took 1.34 seconds to deallocate network for instance. [ 789.950984] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.951580] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.954845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.850s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.955035] env[62066]: DEBUG nova.objects.instance [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lazy-loading 'resources' on Instance uuid 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.037103] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updated VIF entry in instance network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.037645] env[62066]: DEBUG nova.network.neutron [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.056376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.132423] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156047, 'name': CreateVM_Task, 'duration_secs': 0.387351} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.132620] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.133427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.133582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.134265] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.134341] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81ec1599-9728-44ca-848d-7a6400babe6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.138903] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 790.138903] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f2ee92-0611-fc69-b7ff-23379c30a20c" [ 790.138903] env[62066]: _type = "Task" [ 790.138903] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.146510] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f2ee92-0611-fc69-b7ff-23379c30a20c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.308091] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.778385} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.311834] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 43d64f76-7f7a-4b95-b9df-c95218612998/43d64f76-7f7a-4b95-b9df-c95218612998.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 790.312096] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.312357] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d3756fe-addf-4d3c-b276-1503ab8eef1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.318911] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156048, 'name': CreateVM_Task, 'duration_secs': 0.365477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.319964] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.320317] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 790.320317] env[62066]: value = "task-1156049" [ 790.320317] env[62066]: _type = "Task" [ 790.320317] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.322730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.332270] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.349581] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.437628] env[62066]: DEBUG nova.network.neutron [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updated VIF entry in instance network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.438018] env[62066]: DEBUG nova.network.neutron [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.461373] env[62066]: DEBUG nova.compute.utils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 790.463502] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.464357] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.485012] env[62066]: DEBUG nova.compute.manager [req-b547bdad-e8c2-443e-a10d-7f6a412f2280 req-d2c72036-c531-4152-9f19-6a2cdf7614dc service nova] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Received event network-vif-deleted-4627c549-cbf1-40f9-94bb-0f08255025ae {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.509428] env[62066]: DEBUG nova.policy [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85a0637645c443a9d9e6028053c0450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '350ba3c5676a4dd0a018900e7237a5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.541132] env[62066]: DEBUG oslo_concurrency.lockutils [req-d4185043-53af-48be-be7e-6ce691adf71d req-20a04566-01fd-4c5a-8866-27c5d425f447 service nova] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.541588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.650802] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f2ee92-0611-fc69-b7ff-23379c30a20c, 'name': SearchDatastore_Task, 'duration_secs': 0.01065} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.651272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.651581] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.651887] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.652097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.652333] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.652672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.653039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.653342] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f3e43ac-ac1e-4f7a-b968-207d9e5ded1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.655377] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5f6c28-9611-4446-a497-202500208d9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.668313] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 790.668313] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526a0963-2864-7865-e101-9770fa1ea4dc" [ 790.668313] env[62066]: _type = "Task" [ 790.668313] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.672724] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.673088] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.677021] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5930c0d-7bd4-404f-bbd0-1e4491669633 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.682283] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526a0963-2864-7865-e101-9770fa1ea4dc, 'name': SearchDatastore_Task, 'duration_secs': 0.011014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.683540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.683946] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.684420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.688337] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 790.688337] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ded88d-2260-9dd3-2aa2-b7717872327e" [ 790.688337] env[62066]: _type = "Task" [ 790.688337] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.696726] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ded88d-2260-9dd3-2aa2-b7717872327e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.835014] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064094} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.838858] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.840518] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c527a5-0f34-49eb-a965-05cb0f4355eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.868850] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 43d64f76-7f7a-4b95-b9df-c95218612998/43d64f76-7f7a-4b95-b9df-c95218612998.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.873237] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc43d24f-bba9-4a70-82d1-fb56806948b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.897021] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 790.897021] env[62066]: value = "task-1156050" [ 790.897021] env[62066]: _type = "Task" [ 790.897021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.906361] env[62066]: DEBUG nova.compute.manager [None req-48a81793-918a-4332-a9f7-a9679c7e3a7f tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.907187] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.908111] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607e6721-7522-451b-ae96-dae199f2b15e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.941482] env[62066]: DEBUG oslo_concurrency.lockutils [req-c47dfbc1-c2c2-42b8-8ff4-0061d9ceae1a req-3b9d9e5f-819c-4567-a721-73705a43b924 service nova] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.967393] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.990288] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3298f8df-492c-4778-a6ae-efcb42454685 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.001461] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8d1c0a-ad66-46c4-b1bb-4b015d8bb563 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.046037] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Successfully created port: 9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.048293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cda78b3-e968-41f1-8318-001915771bf9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.059126] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18cbedd-dbfb-463c-bd2b-8bc94ef0b101 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.074117] env[62066]: DEBUG nova.compute.provider_tree [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.082503] env[62066]: DEBUG nova.network.neutron [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.146414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.146725] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.199639] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ded88d-2260-9dd3-2aa2-b7717872327e, 'name': SearchDatastore_Task, 'duration_secs': 0.01216} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.200551] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60bcbd4f-bb16-425d-b798-951362ff4acc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.205993] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 791.205993] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c884c7-77e7-2f95-ef7e-9cd054de7956" [ 791.205993] env[62066]: _type = "Task" [ 791.205993] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.214244] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c884c7-77e7-2f95-ef7e-9cd054de7956, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.274655] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.274746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.275582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.275582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.275582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.277742] env[62066]: INFO nova.compute.manager [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Terminating instance [ 791.281359] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "refresh_cache-e95ee795-603e-4cbf-bcd6-1ba54b62a281" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.281439] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquired lock "refresh_cache-e95ee795-603e-4cbf-bcd6-1ba54b62a281" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.281609] env[62066]: DEBUG nova.network.neutron [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.406209] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.418353] env[62066]: INFO nova.compute.manager [None req-48a81793-918a-4332-a9f7-a9679c7e3a7f tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] instance snapshotting [ 791.419136] env[62066]: DEBUG nova.objects.instance [None req-48a81793-918a-4332-a9f7-a9679c7e3a7f tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lazy-loading 'flavor' on Instance uuid e95ee795-603e-4cbf-bcd6-1ba54b62a281 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.578582] env[62066]: DEBUG nova.scheduler.client.report [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.650678] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.722542] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c884c7-77e7-2f95-ef7e-9cd054de7956, 'name': SearchDatastore_Task, 'duration_secs': 0.048715} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.728032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.728032] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4723be94-3479-4e66-8088-914824c0e669/4723be94-3479-4e66-8088-914824c0e669.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.728032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.728032] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.728032] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c9687af-ac4a-4af3-9543-99d2e8f253da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.728938] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c52df10b-36f7-468f-9a52-4d55f0c04d10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.736835] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 791.736835] env[62066]: value = "task-1156051" [ 791.736835] env[62066]: _type = "Task" [ 791.736835] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.740584] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.740776] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.745022] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e6903b4-e46b-43ce-ba91-8202e9608b47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.747678] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156051, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.751387] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 791.751387] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528074b1-1188-ed38-f82e-55c5143d0462" [ 791.751387] env[62066]: _type = "Task" [ 791.751387] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.764712] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528074b1-1188-ed38-f82e-55c5143d0462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.813382] env[62066]: DEBUG nova.network.neutron [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.833254] env[62066]: DEBUG nova.network.neutron [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.878246] env[62066]: DEBUG nova.network.neutron [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.906603] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156050, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.917651] env[62066]: DEBUG nova.compute.manager [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.919245] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9271dba5-ee2d-44af-924a-4879ecb6205d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.926052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93db4bea-381d-4e50-8b8b-c01692b4f97c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.945455] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcc0109-177f-4085-b588-00025edcf12f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.963241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-8a5014e0-3b5b-46a1-b1b6-b98e168c945f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.963241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-8a5014e0-3b5b-46a1-b1b6-b98e168c945f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.978476] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.020835] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.021145] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.021338] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.021538] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.021692] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.021849] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.022197] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.022428] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.022722] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.022922] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.023138] env[62066]: DEBUG nova.virt.hardware [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.024071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7148817-0d8b-4fbc-a387-cd1412c44e1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.035177] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc04e0a-8cfb-4882-a71f-57ae13a23c19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.090658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.134s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.092510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.763s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.094122] env[62066]: INFO nova.compute.claims [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.164079] env[62066]: INFO nova.scheduler.client.report [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Deleted allocations for instance 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2 [ 792.182890] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.246874] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156051, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.261964] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528074b1-1188-ed38-f82e-55c5143d0462, 'name': SearchDatastore_Task, 'duration_secs': 0.024675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.262811] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c2b9d2b-9710-4d34-a2e9-8ff6640bebbe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.269155] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 792.269155] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ff6965-e780-bb53-5a1e-d8f660fecc97" [ 792.269155] env[62066]: _type = "Task" [ 792.269155] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.277944] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ff6965-e780-bb53-5a1e-d8f660fecc97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.337809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.337809] env[62066]: DEBUG nova.compute.manager [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Inject network info {{(pid=62066) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 792.338023] env[62066]: DEBUG nova.compute.manager [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] network_info to inject: |[{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 792.343727] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfiguring VM instance to set the machine id {{(pid=62066) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 792.343727] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2226c5bf-f1e3-4b0e-8f3a-9e7be7db29e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.358633] env[62066]: DEBUG oslo_vmware.api [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 792.358633] env[62066]: value = "task-1156052" [ 792.358633] env[62066]: _type = "Task" [ 792.358633] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.369145] env[62066]: DEBUG oslo_vmware.api [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.383198] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Releasing lock "refresh_cache-e95ee795-603e-4cbf-bcd6-1ba54b62a281" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.383654] env[62066]: DEBUG nova.compute.manager [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.384214] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.384900] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ebfdcc-f9af-4691-a087-688a0638e012 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.392945] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 792.393278] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66cf5d14-f50e-461f-bed1-015ce795716c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.402371] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 792.402371] env[62066]: value = "task-1156053" [ 792.402371] env[62066]: _type = "Task" [ 792.402371] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.409556] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156050, 'name': ReconfigVM_Task, 'duration_secs': 1.076151} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.410368] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 43d64f76-7f7a-4b95-b9df-c95218612998/43d64f76-7f7a-4b95-b9df-c95218612998.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.411049] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d600ec0c-36ba-4bf3-862d-ce887f39c94c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.418657] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.420283] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 792.420283] env[62066]: value = "task-1156054" [ 792.420283] env[62066]: _type = "Task" [ 792.420283] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.430415] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156054, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.434836] env[62066]: INFO nova.compute.manager [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] instance snapshotting [ 792.435503] env[62066]: DEBUG nova.objects.instance [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.457287] env[62066]: DEBUG nova.compute.manager [None req-48a81793-918a-4332-a9f7-a9679c7e3a7f tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance disappeared during snapshot {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 792.466501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.466617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.467812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536a0c95-36fd-496e-999c-e44e39b2cc3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.493072] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ad6b4a-9009-428c-8b44-3ef7cedaf9dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.522054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 792.526248] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e0fe36a-8bb5-4efe-bd22-537aba8cfb8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.545929] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 792.545929] env[62066]: value = "task-1156055" [ 792.545929] env[62066]: _type = "Task" [ 792.545929] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.556650] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.589369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.589650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.589835] env[62066]: INFO nova.compute.manager [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Shelving [ 792.641485] env[62066]: DEBUG nova.compute.manager [None req-48a81793-918a-4332-a9f7-a9679c7e3a7f tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Found 0 images (rotation: 2) {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 792.641764] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Successfully updated port: 9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.645203] env[62066]: DEBUG nova.compute.manager [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.645405] env[62066]: DEBUG nova.compute.manager [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing instance network info cache due to event network-changed-e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.645619] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] Acquiring lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.645759] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] Acquired lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.645930] env[62066]: DEBUG nova.network.neutron [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Refreshing network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.677025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e93ef06e-2252-4cda-b8a1-90c9576b43c3 tempest-InstanceActionsTestJSON-1193526571 tempest-InstanceActionsTestJSON-1193526571-project-member] Lock "16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.765s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.746508] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156051, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.755809} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.746786] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4723be94-3479-4e66-8088-914824c0e669/4723be94-3479-4e66-8088-914824c0e669.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.747019] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.747271] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-071f6407-b5b5-402e-b2cd-8817df5057dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.754708] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 792.754708] env[62066]: value = "task-1156056" [ 792.754708] env[62066]: _type = "Task" [ 792.754708] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.764748] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156056, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.781268] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ff6965-e780-bb53-5a1e-d8f660fecc97, 'name': SearchDatastore_Task, 'duration_secs': 0.064542} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.781555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.781817] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.782096] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7a92ad7-3092-42c3-a9fe-8185ffe6a6be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.788977] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 792.788977] env[62066]: value = "task-1156057" [ 792.788977] env[62066]: _type = "Task" [ 792.788977] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.798389] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.868048] env[62066]: DEBUG oslo_vmware.api [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156052, 'name': ReconfigVM_Task, 'duration_secs': 0.167917} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.868367] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8263755b-4445-4fa6-8291-99cbecf68207 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Reconfigured VM instance to set the machine id {{(pid=62066) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 792.912880] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156053, 'name': PowerOffVM_Task, 'duration_secs': 0.15974} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.913166] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 792.913385] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 792.913581] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0784c347-42c3-460a-afa9-fb47e5af31a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.929584] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156054, 'name': Rename_Task, 'duration_secs': 0.270648} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.930203] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.930466] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51b61276-d9cf-4212-9b23-1965f24ec40f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.936249] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 792.936508] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 792.936778] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Deleting the datastore file [datastore2] e95ee795-603e-4cbf-bcd6-1ba54b62a281 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.938121] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-994474e8-ea81-4e35-b8cc-cf2e46ca78b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.944845] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 792.944845] env[62066]: value = "task-1156059" [ 792.944845] env[62066]: _type = "Task" [ 792.944845] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.951592] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for the task: (returnval){ [ 792.951592] env[62066]: value = "task-1156060" [ 792.951592] env[62066]: _type = "Task" [ 792.951592] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.953542] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d039dc5-1f0c-4402-a8ea-dd1880bb23fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.962890] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156059, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.968713] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.982647] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5627260-5e66-49c1-8e3a-5716f807ca2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.056960] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.098187] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.098672] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0bc6570-1bd0-4214-b2c1-f17215671199 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.106809] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 793.106809] env[62066]: value = "task-1156061" [ 793.106809] env[62066]: _type = "Task" [ 793.106809] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.119159] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.148755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.148918] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.149087] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.227065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "2f186d0a-91a0-4dc2-83bd-511099445af7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.227347] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.227741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.227991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.228338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.235177] env[62066]: INFO nova.compute.manager [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Terminating instance [ 793.237491] env[62066]: DEBUG nova.compute.manager [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.237725] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.238656] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c20613-9ac8-4166-b20c-2f1fdae0150d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.248139] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.250735] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-775a9d61-1612-43ba-b6bb-f77c0c85b1a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.257843] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 793.257843] env[62066]: value = "task-1156062" [ 793.257843] env[62066]: _type = "Task" [ 793.257843] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.275956] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156056, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.280020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.280020] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.283139] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5832afe-2fa5-4349-9872-c76edab1ec18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.310295] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 4723be94-3479-4e66-8088-914824c0e669/4723be94-3479-4e66-8088-914824c0e669.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.317105] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0570131b-de18-4570-ab3d-5e4398cc36d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.337531] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156057, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.341632] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 793.341632] env[62066]: value = "task-1156063" [ 793.341632] env[62066]: _type = "Task" [ 793.341632] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.353872] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156063, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.461892] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156059, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.471487] env[62066]: DEBUG oslo_vmware.api [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Task: {'id': task-1156060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286106} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.471487] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.471487] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 793.471487] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.471487] env[62066]: INFO nova.compute.manager [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Took 1.09 seconds to destroy the instance on the hypervisor. [ 793.471833] env[62066]: DEBUG oslo.service.loopingcall [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.471833] env[62066]: DEBUG nova.compute.manager [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.471833] env[62066]: DEBUG nova.network.neutron [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 793.493991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 793.494151] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-84caffa6-df40-45fa-957c-f059b5aad2eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.500584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.500854] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.501177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.501377] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.501554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.505858] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 793.505858] env[62066]: value = "task-1156064" [ 793.505858] env[62066]: _type = "Task" [ 793.505858] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.509173] env[62066]: INFO nova.compute.manager [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Terminating instance [ 793.511390] env[62066]: DEBUG nova.network.neutron [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.513706] env[62066]: DEBUG nova.compute.manager [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.513921] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.514882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0dd26b-b303-494e-8bfd-613841ce3b1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.522221] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.534028] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.534645] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55786ebe-f0bd-47cb-ab97-bf7c62d454bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.542031] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 793.542031] env[62066]: value = "task-1156065" [ 793.542031] env[62066]: _type = "Task" [ 793.542031] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.560565] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.563827] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.605818] env[62066]: DEBUG nova.network.neutron [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updated VIF entry in instance network info cache for port e8c586f5-3024-40d6-934b-04b551be3c44. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.606280] env[62066]: DEBUG nova.network.neutron [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [{"id": "e8c586f5-3024-40d6-934b-04b551be3c44", "address": "fa:16:3e:47:b9:5f", "network": {"id": "5688ca15-b7b6-41a4-9a5e-96f864cd5e56", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1374253263-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70d709f95cfa418392c56eb730c053c3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8c586f5-30", "ovs_interfaceid": "e8c586f5-3024-40d6-934b-04b551be3c44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.620439] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156061, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.658802] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a703e9b5-0149-4407-a56a-95510cb46bee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.667153] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9f5390-237d-4f05-9185-b078aaa22789 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.701206] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.703856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dde8e15-b607-4dc8-8d1b-3c4891ed43ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.712546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f1bf09-e28e-48dc-9ff3-ea05b442c22f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.726895] env[62066]: DEBUG nova.compute.provider_tree [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.776283] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156062, 'name': PowerOffVM_Task, 'duration_secs': 0.232014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.776623] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 793.776816] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.777095] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7eb3682-08ad-4df9-8972-cd6227f356a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.810300] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.879173} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.810608] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.810836] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.811110] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-26b10d70-dad9-43d9-b942-ee9df3bf1a46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.819227] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 793.819227] env[62066]: value = "task-1156067" [ 793.819227] env[62066]: _type = "Task" [ 793.819227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.828840] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.838632] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.838878] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.839083] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleting the datastore file [datastore2] 2f186d0a-91a0-4dc2-83bd-511099445af7 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.839382] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af074c56-1348-4bdd-8cab-90bafd1cb180 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.847400] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 793.847400] env[62066]: value = "task-1156068" [ 793.847400] env[62066]: _type = "Task" [ 793.847400] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.856210] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156063, 'name': ReconfigVM_Task, 'duration_secs': 0.446291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.856959] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 4723be94-3479-4e66-8088-914824c0e669/4723be94-3479-4e66-8088-914824c0e669.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.857720] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df4b4287-b8a5-4372-a348-deb20a4462b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.862608] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156068, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.867224] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 793.867224] env[62066]: value = "task-1156069" [ 793.867224] env[62066]: _type = "Task" [ 793.867224] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.876728] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156069, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.943875] env[62066]: DEBUG nova.network.neutron [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Updating instance_info_cache with network_info: [{"id": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "address": "fa:16:3e:0c:3f:cd", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c31bd1b-5e", "ovs_interfaceid": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.955024] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156059, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.014577] env[62066]: DEBUG nova.network.neutron [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.015884] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156064, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.058448] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156065, 'name': PowerOffVM_Task, 'duration_secs': 0.198607} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.059015] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.059268] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 794.059753] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2514a670-98f5-4831-8e1f-13a7669cc087 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.065958] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.113939] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f101651-565f-4937-8808-d5572889951c req-ce75277a-4183-4e83-981f-915854a3c35d service nova] Releasing lock "refresh_cache-82e2a71a-d27f-4db9-8f84-16762d3d3bf6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.120618] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156061, 'name': PowerOffVM_Task, 'duration_secs': 0.885058} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.121632] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.123438] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b07cf9-1197-4b5f-8ecc-e1b39739d75b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.126973] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.127342] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.127588] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Deleting the datastore file [datastore1] 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.129060] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8caef476-92d2-4b27-9686-58e4ea54711c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.148847] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f4f1b1-da57-4d54-b9b3-3902add66eb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.151956] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for the task: (returnval){ [ 794.151956] env[62066]: value = "task-1156071" [ 794.151956] env[62066]: _type = "Task" [ 794.151956] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.165541] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.232655] env[62066]: DEBUG nova.scheduler.client.report [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.332373] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06892} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.332453] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.333430] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b289a758-467e-4aa4-b7a3-bfa1afa87fd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.359597] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.363461] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c9e5a23-c382-4316-a055-22018826ebcf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.391495] env[62066]: DEBUG oslo_vmware.api [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156068, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326913} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.392562] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.392830] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.393083] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.393448] env[62066]: INFO nova.compute.manager [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 794.393968] env[62066]: DEBUG oslo.service.loopingcall [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.398016] env[62066]: DEBUG nova.compute.manager [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.398163] env[62066]: DEBUG nova.network.neutron [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.400415] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 794.400415] env[62066]: value = "task-1156072" [ 794.400415] env[62066]: _type = "Task" [ 794.400415] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.401065] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156069, 'name': Rename_Task, 'duration_secs': 0.158531} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.401474] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.405242] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f900adb0-e8c7-4335-9c76-648cbb47622b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.414625] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156072, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.416642] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 794.416642] env[62066]: value = "task-1156073" [ 794.416642] env[62066]: _type = "Task" [ 794.416642] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.425274] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156073, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.447201] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.447522] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance network_info: |[{"id": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "address": "fa:16:3e:0c:3f:cd", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c31bd1b-5e", "ovs_interfaceid": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.448112] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:3f:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.458630] env[62066]: DEBUG oslo.service.loopingcall [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.462628] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.463024] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9ae5f47-1b2c-4c24-8609-d15168b0573e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.485600] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156059, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.486914] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.486914] env[62066]: value = "task-1156074" [ 794.486914] env[62066]: _type = "Task" [ 794.486914] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.495666] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156074, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.517557] env[62066]: INFO nova.compute.manager [-] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Took 1.05 seconds to deallocate network for instance. [ 794.526813] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156064, 'name': CreateSnapshot_Task, 'duration_secs': 0.677659} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.527258] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 794.528372] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a4c442-f067-4746-9496-2ccf295f9bc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.566148] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.664731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 794.665449] env[62066]: DEBUG oslo_vmware.api [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Task: {'id': task-1156071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.401004} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.665795] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d12693f0-9104-4b6c-8a56-631beff3d2f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.669593] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.670456] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.670550] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.670779] env[62066]: INFO nova.compute.manager [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 794.671013] env[62066]: DEBUG oslo.service.loopingcall [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.672059] env[62066]: DEBUG nova.compute.manager [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.672059] env[62066]: DEBUG nova.network.neutron [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.679092] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 794.679092] env[62066]: value = "task-1156075" [ 794.679092] env[62066]: _type = "Task" [ 794.679092] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.688571] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156075, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.737791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.738251] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 794.741298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 28.535s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.741479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.741632] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 794.741922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.844s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.743513] env[62066]: INFO nova.compute.claims [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.748525] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f87012-9ada-4697-be37-29191634e249 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.761773] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a6539c-04da-41b8-b011-461b62e97dfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.783435] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0f3d7e-d68a-4f16-9058-4f77fb3f71ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.795199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3a65a1-a7ca-4288-9d4c-4c73fd4ebaaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.834809] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180187MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 794.834809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.837814] env[62066]: DEBUG nova.compute.manager [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Received event network-vif-plugged-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.837949] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Acquiring lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.841307] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.841307] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.841307] env[62066]: DEBUG nova.compute.manager [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] No waiting events found dispatching network-vif-plugged-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.841307] env[62066]: WARNING nova.compute.manager [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Received unexpected event network-vif-plugged-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf for instance with vm_state building and task_state spawning. [ 794.841307] env[62066]: DEBUG nova.compute.manager [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Received event network-changed-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.841307] env[62066]: DEBUG nova.compute.manager [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Refreshing instance network info cache due to event network-changed-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.841307] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Acquiring lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.841307] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Acquired lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.841307] env[62066]: DEBUG nova.network.neutron [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Refreshing network info cache for port 9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.916869] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156072, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.928707] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156073, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.957325] env[62066]: DEBUG oslo_vmware.api [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156059, 'name': PowerOnVM_Task, 'duration_secs': 1.894316} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.957615] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.957826] env[62066]: INFO nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Took 10.31 seconds to spawn the instance on the hypervisor. [ 794.958043] env[62066]: DEBUG nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.958874] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3dc2d0-fcea-4e68-a324-6f36e2d0cd2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.003558] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156074, 'name': CreateVM_Task, 'duration_secs': 0.492745} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.003558] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.003558] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.003558] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.003558] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.003558] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6f3bc07-fe8d-4d61-9ca8-cff32a068ffb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.008418] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 795.008418] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b08b78-80e1-3bf4-9c23-69ff23b80baa" [ 795.008418] env[62066]: _type = "Task" [ 795.008418] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.019164] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b08b78-80e1-3bf4-9c23-69ff23b80baa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.026097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.048574] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 795.049364] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6a094622-522d-4bff-8c60-11be8fbbf5fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.064056] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.065873] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 795.065873] env[62066]: value = "task-1156076" [ 795.065873] env[62066]: _type = "Task" [ 795.065873] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.074796] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156076, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.191643] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156075, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.249530] env[62066]: DEBUG nova.compute.utils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 795.252354] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.252593] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 795.330801] env[62066]: DEBUG nova.policy [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '315ec256cd6b422a90f2914175cb49bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c93c0e92cfec42f4b4a20e9fb4a32088', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.347406] env[62066]: DEBUG nova.network.neutron [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.413563] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156072, 'name': ReconfigVM_Task, 'duration_secs': 0.609504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.413895] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Reconfigured VM instance instance-0000003d to attach disk [datastore1] d9163e64-309b-4381-8819-15757f83ac2e/d9163e64-309b-4381-8819-15757f83ac2e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.414527] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de688bdc-60a7-4bfe-980c-f082eb448cef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.421916] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 795.421916] env[62066]: value = "task-1156077" [ 795.421916] env[62066]: _type = "Task" [ 795.421916] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.429218] env[62066]: DEBUG oslo_vmware.api [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156073, 'name': PowerOnVM_Task, 'duration_secs': 0.635082} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.429860] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.430135] env[62066]: INFO nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Took 8.13 seconds to spawn the instance on the hypervisor. [ 795.430425] env[62066]: DEBUG nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.431111] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2304956-91db-4ea3-9a92-643fefa71da1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.438222] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156077, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.483869] env[62066]: INFO nova.compute.manager [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Took 35.41 seconds to build instance. [ 795.522421] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b08b78-80e1-3bf4-9c23-69ff23b80baa, 'name': SearchDatastore_Task, 'duration_secs': 0.01905} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.522779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.523034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.523275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.523466] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.523725] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.524603] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fce2dc2-ed6a-4ca9-a93b-cbe739df2dc4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.535051] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.535558] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 795.536564] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3db8098d-0ccb-4271-bb9d-666ec7dbb5c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.543024] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 795.543024] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5241121d-5865-3da5-302a-ffc6ba658869" [ 795.543024] env[62066]: _type = "Task" [ 795.543024] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.552213] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5241121d-5865-3da5-302a-ffc6ba658869, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.562807] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.580335] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156076, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.700412] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156075, 'name': CreateSnapshot_Task, 'duration_secs': 0.741049} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.700966] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 795.701625] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8a903a-ab4f-46ae-9c7c-a01884944364 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.753232] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.849977] env[62066]: INFO nova.compute.manager [-] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Took 1.45 seconds to deallocate network for instance. [ 795.911303] env[62066]: DEBUG nova.network.neutron [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Updated VIF entry in instance network info cache for port 9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 795.911927] env[62066]: DEBUG nova.network.neutron [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Updating instance_info_cache with network_info: [{"id": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "address": "fa:16:3e:0c:3f:cd", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c31bd1b-5e", "ovs_interfaceid": "9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.935881] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156077, 'name': Rename_Task, 'duration_secs': 0.180081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.937785] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.937963] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9af48199-1297-4c33-9850-693eb3ef9274 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.946306] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 795.946306] env[62066]: value = "task-1156078" [ 795.946306] env[62066]: _type = "Task" [ 795.946306] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.965761] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156078, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.966319] env[62066]: INFO nova.compute.manager [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Took 34.36 seconds to build instance. [ 795.987507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6310918-4ee1-435d-9a6c-a0831b9b7e6c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.053s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.057983] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5241121d-5865-3da5-302a-ffc6ba658869, 'name': SearchDatastore_Task, 'duration_secs': 0.013339} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.062737] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35bc5888-efdd-489c-b4cf-249078ea196d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.079783] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.080281] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 796.080281] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529e19aa-6190-380c-6535-b4fd0d6c2761" [ 796.080281] env[62066]: _type = "Task" [ 796.080281] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.085889] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Successfully created port: 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.092890] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156076, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.103684] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529e19aa-6190-380c-6535-b4fd0d6c2761, 'name': SearchDatastore_Task, 'duration_secs': 0.010895} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.104647] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.104647] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.105196] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-457271e0-e735-4e40-b6f8-33df45d80015 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.113777] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 796.113777] env[62066]: value = "task-1156079" [ 796.113777] env[62066]: _type = "Task" [ 796.113777] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.127803] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.224056] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 796.227984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-435b2732-75e5-4415-9c66-3c22929af6d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.237848] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 796.237848] env[62066]: value = "task-1156080" [ 796.237848] env[62066]: _type = "Task" [ 796.237848] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.250662] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.280014] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1733c92-74ad-409d-b090-eb1fd4cab02d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.289465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62106330-c928-4233-9a7a-2bd7ba13b534 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.332407] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6d7c82-293f-494c-bc8d-ccef905335a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.348881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6200a95e-c7a9-45e6-9305-5695ce922f97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.368163] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.369068] env[62066]: DEBUG nova.compute.provider_tree [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.416591] env[62066]: DEBUG oslo_concurrency.lockutils [req-0fcd0ef7-2e8e-42be-baa8-e18065fc21be req-c0367eab-dd59-4165-9015-e6e221ae11c0 service nova] Releasing lock "refresh_cache-b46b7795-b7f7-4c55-bc72-184bb7bb8842" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.458305] env[62066]: DEBUG oslo_vmware.api [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156078, 'name': PowerOnVM_Task, 'duration_secs': 0.480642} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.458785] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.459121] env[62066]: DEBUG nova.compute.manager [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.460043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc24e6b-3490-44b3-bf75-be9b6b1ac611 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.470404] env[62066]: DEBUG oslo_concurrency.lockutils [None req-38c06401-22ff-4f11-a95c-223b2e9768ed tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.207s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.571437] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.587865] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156076, 'name': CloneVM_Task} progress is 95%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.630274] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156079, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.652914] env[62066]: DEBUG nova.compute.manager [req-e7ef1824-79bc-4c89-bafe-2eb15148b379 req-3ee99ee8-ec5c-47ae-b406-55ae9ab49ba4 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Received event network-vif-deleted-e8c586f5-3024-40d6-934b-04b551be3c44 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.653312] env[62066]: INFO nova.compute.manager [req-e7ef1824-79bc-4c89-bafe-2eb15148b379 req-3ee99ee8-ec5c-47ae-b406-55ae9ab49ba4 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Neutron deleted interface e8c586f5-3024-40d6-934b-04b551be3c44; detaching it from the instance and deleting it from the info cache [ 796.653587] env[62066]: DEBUG nova.network.neutron [req-e7ef1824-79bc-4c89-bafe-2eb15148b379 req-3ee99ee8-ec5c-47ae-b406-55ae9ab49ba4 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.747632] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.770050] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.803037] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.807020] env[62066]: DEBUG nova.virt.hardware [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.807020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb3daf4-0220-44e2-8a85-ca52b7843188 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.819633] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e900f5f5-5b20-4cdb-869d-fce9daf8f6f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.836889] env[62066]: DEBUG nova.network.neutron [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.842395] env[62066]: INFO nova.compute.manager [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Rescuing [ 796.842395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.842671] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.842671] env[62066]: DEBUG nova.network.neutron [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.872643] env[62066]: DEBUG nova.scheduler.client.report [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.985152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.071848] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.083774] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156076, 'name': CloneVM_Task, 'duration_secs': 1.768227} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.084279] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created linked-clone VM from snapshot [ 797.085598] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2fe8b3-789d-4797-bd1c-3583c8656bfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.098020] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploading image 59e9c507-e2ea-4178-b84f-f16bfba53d2a {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 797.133349] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156079, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612797} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.135661] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 797.135661] env[62066]: value = "vm-251679" [ 797.135661] env[62066]: _type = "VirtualMachine" [ 797.135661] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 797.136183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.136301] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.136798] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-16a82a49-8c63-40b3-924c-3a49df20b22e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.138209] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae5910f4-8af3-4210-b076-1a6ce46acaed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.145210] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 797.145210] env[62066]: value = "task-1156081" [ 797.145210] env[62066]: _type = "Task" [ 797.145210] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.146624] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease: (returnval){ [ 797.146624] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52036280-1022-237c-b747-83bb8225acee" [ 797.146624] env[62066]: _type = "HttpNfcLease" [ 797.146624] env[62066]: } obtained for exporting VM: (result){ [ 797.146624] env[62066]: value = "vm-251679" [ 797.146624] env[62066]: _type = "VirtualMachine" [ 797.146624] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 797.147025] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the lease: (returnval){ [ 797.147025] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52036280-1022-237c-b747-83bb8225acee" [ 797.147025] env[62066]: _type = "HttpNfcLease" [ 797.147025] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 797.160581] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.162556] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4db30a14-b938-4659-b5e9-786d97b4a111 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.164454] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 797.164454] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52036280-1022-237c-b747-83bb8225acee" [ 797.164454] env[62066]: _type = "HttpNfcLease" [ 797.164454] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 797.171597] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef45240-880e-485f-9244-826c75f1c697 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.205507] env[62066]: DEBUG nova.compute.manager [req-e7ef1824-79bc-4c89-bafe-2eb15148b379 req-3ee99ee8-ec5c-47ae-b406-55ae9ab49ba4 service nova] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Detach interface failed, port_id=e8c586f5-3024-40d6-934b-04b551be3c44, reason: Instance 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 797.207889] env[62066]: DEBUG nova.compute.manager [req-eff4e15b-1bce-4d85-a87d-0846d5f8231a req-5e498d19-d95a-4e42-88d9-d043b72e2f16 service nova] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Received event network-vif-deleted-1cfa47b1-d2b5-4417-9791-d751e1c777fe {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.249789] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.339727] env[62066]: INFO nova.compute.manager [-] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Took 2.67 seconds to deallocate network for instance. [ 797.381484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.382381] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.385095] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.481s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.385446] env[62066]: DEBUG nova.objects.instance [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lazy-loading 'resources' on Instance uuid 1c694323-616c-43b1-bd49-bba707e0788f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 797.585027] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.624223] env[62066]: DEBUG nova.network.neutron [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.662456] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069967} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.663582] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.664356] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08ce745-2307-4815-942a-81492c787e21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.674410] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 797.674410] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52036280-1022-237c-b747-83bb8225acee" [ 797.674410] env[62066]: _type = "HttpNfcLease" [ 797.674410] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 797.674410] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 797.674410] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52036280-1022-237c-b747-83bb8225acee" [ 797.674410] env[62066]: _type = "HttpNfcLease" [ 797.674410] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 797.674410] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f811d7b3-5bd3-4e79-a1bb-7a78d65e4a32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.698107] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.701723] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e69557ed-c07d-46d3-873e-693e5d5fa930 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.728363] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 797.728448] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 797.793726] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 797.793726] env[62066]: value = "task-1156083" [ 797.793726] env[62066]: _type = "Task" [ 797.793726] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.803282] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.808950] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156083, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.845550] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e8880797-3090-4b86-9929-db0795f2ff4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.851118] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.891156] env[62066]: DEBUG nova.compute.utils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 797.901143] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.901143] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 798.010126] env[62066]: DEBUG nova.policy [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e36a41f0ee3a4bd5815e5067b4114ac4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00dcb6ace2714cd4994a5273586c5b95', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.077707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "43d64f76-7f7a-4b95-b9df-c95218612998" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.077983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.078599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.078812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.079117] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.081253] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.081716] env[62066]: INFO nova.compute.manager [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Terminating instance [ 798.083845] env[62066]: DEBUG nova.compute.manager [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.085091] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 798.085896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516ffd8d-aac9-4b2e-a6d5-a3db68a455be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.093736] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.094271] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-071ae1a0-a1da-4a42-8a9e-0fc05a3a510b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.103587] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 798.103587] env[62066]: value = "task-1156084" [ 798.103587] env[62066]: _type = "Task" [ 798.103587] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.115760] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.126640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.135673] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Successfully updated port: 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.231732] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "d9163e64-309b-4381-8819-15757f83ac2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.232078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.232411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "d9163e64-309b-4381-8819-15757f83ac2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.232638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.233081] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.236804] env[62066]: INFO nova.compute.manager [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Terminating instance [ 798.240472] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "refresh_cache-d9163e64-309b-4381-8819-15757f83ac2e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.240660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquired lock "refresh_cache-d9163e64-309b-4381-8819-15757f83ac2e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.240948] env[62066]: DEBUG nova.network.neutron [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.308660] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task} progress is 95%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.324539] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156083, 'name': ReconfigVM_Task, 'duration_secs': 0.485553} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.325016] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Reconfigured VM instance instance-00000042 to attach disk [datastore2] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.326974] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1eca0784-8d18-4d3c-a307-9c0d889cd4a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.340331] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 798.340331] env[62066]: value = "task-1156085" [ 798.340331] env[62066]: _type = "Task" [ 798.340331] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.355543] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156085, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.405249] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.579426] env[62066]: DEBUG oslo_vmware.api [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156055, 'name': ReconfigVM_Task, 'duration_secs': 5.799675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.579426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.579426] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 798.587766] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f935ea98-4bca-4896-bd50-fd50489e49a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.596932] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6295f500-bd35-484d-8bdd-f3363d78bdee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.637274] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c109511f-ac71-49cc-8a12-62f64b2c6038 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.640676] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.640882] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.641066] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.648088] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156084, 'name': PowerOffVM_Task, 'duration_secs': 0.244704} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.648898] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.649574] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.650958] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1090ca0-5265-4f40-82bc-d239164304c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.655783] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05410b36-9e45-4e20-ac2e-eecd9b28e007 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.668839] env[62066]: DEBUG nova.compute.provider_tree [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.689948] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.690807] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e47da86-af84-4270-8367-c7138d97c86d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.700346] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 798.700346] env[62066]: value = "task-1156087" [ 798.700346] env[62066]: _type = "Task" [ 798.700346] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.709548] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.719967] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.720283] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.720509] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore1] 43d64f76-7f7a-4b95-b9df-c95218612998 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.721269] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f37038c3-a832-458c-921a-4d159bd43459 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.728817] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 798.728817] env[62066]: value = "task-1156088" [ 798.728817] env[62066]: _type = "Task" [ 798.728817] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.737051] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.770902] env[62066]: DEBUG nova.network.neutron [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.800876] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156080, 'name': CloneVM_Task, 'duration_secs': 2.213279} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.803180] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Created linked-clone VM from snapshot [ 798.806797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a201b241-d236-484d-ad90-16d6b0ff4c62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.811731] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Successfully created port: b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.819031] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Uploading image ca07540e-1efc-45fb-a554-db31a70bc000 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 798.849592] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 798.849592] env[62066]: value = "vm-251681" [ 798.849592] env[62066]: _type = "VirtualMachine" [ 798.849592] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 798.850133] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-58aa0da2-5bed-4965-a5d5-564a8c1dbdd8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.858432] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156085, 'name': Rename_Task, 'duration_secs': 0.176182} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.860120] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 798.860573] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lease: (returnval){ [ 798.860573] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b19f62-242d-578f-17e0-ae59de5f4738" [ 798.860573] env[62066]: _type = "HttpNfcLease" [ 798.860573] env[62066]: } obtained for exporting VM: (result){ [ 798.860573] env[62066]: value = "vm-251681" [ 798.860573] env[62066]: _type = "VirtualMachine" [ 798.860573] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 798.861257] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the lease: (returnval){ [ 798.861257] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b19f62-242d-578f-17e0-ae59de5f4738" [ 798.861257] env[62066]: _type = "HttpNfcLease" [ 798.861257] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 798.861498] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0e7d546-4425-41b4-9818-792bba1931e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.871479] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 798.871479] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b19f62-242d-578f-17e0-ae59de5f4738" [ 798.871479] env[62066]: _type = "HttpNfcLease" [ 798.871479] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 798.872950] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 798.872950] env[62066]: value = "task-1156090" [ 798.872950] env[62066]: _type = "Task" [ 798.872950] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.881775] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.905214] env[62066]: DEBUG nova.network.neutron [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.173623] env[62066]: DEBUG nova.scheduler.client.report [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.202248] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.219494] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156087, 'name': PowerOffVM_Task, 'duration_secs': 0.29772} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.219953] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 799.221177] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488e8155-b0fa-4386-95c1-901c0d923c89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.250448] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ad0e7d-f20e-4caf-a761-dad222cdbc6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.261524] env[62066]: DEBUG oslo_vmware.api [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258197} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.262534] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.262823] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.262991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.263196] env[62066]: INFO nova.compute.manager [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Took 1.18 seconds to destroy the instance on the hypervisor. [ 799.263485] env[62066]: DEBUG oslo.service.loopingcall [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.265764] env[62066]: DEBUG nova.compute.manager [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.266091] env[62066]: DEBUG nova.network.neutron [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.297023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.297023] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a7a247e-db07-4a5f-b37c-d63d55c73242 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.302648] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 799.302648] env[62066]: value = "task-1156091" [ 799.302648] env[62066]: _type = "Task" [ 799.302648] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.317748] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 799.317988] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.318260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.318455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.318717] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.319076] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b4d8506-592e-4452-9e87-0684603dc723 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.327609] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.327886] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.328737] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-007c973a-d908-486b-9b31-2c482c4068a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.334636] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 799.334636] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e0d77f-588a-9bd3-40eb-3f056bfe0a72" [ 799.334636] env[62066]: _type = "Task" [ 799.334636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.342753] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e0d77f-588a-9bd3-40eb-3f056bfe0a72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.376306] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 799.376306] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b19f62-242d-578f-17e0-ae59de5f4738" [ 799.376306] env[62066]: _type = "HttpNfcLease" [ 799.376306] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 799.380096] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 799.380096] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b19f62-242d-578f-17e0-ae59de5f4738" [ 799.380096] env[62066]: _type = "HttpNfcLease" [ 799.380096] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 799.384541] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e092a7b-4f12-45fa-9ffa-b00a4acd630a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.397857] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156090, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.402034] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 799.402454] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 799.465371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Releasing lock "refresh_cache-d9163e64-309b-4381-8819-15757f83ac2e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.465892] env[62066]: DEBUG nova.compute.manager [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 799.467045] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.467555] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.475971] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a189b2-28df-4b0e-b14c-4fe8b65dd0c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.486794] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.486794] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6e9ae50-cd6a-4209-a32d-ef9a28b25c41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.495568] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 799.495568] env[62066]: value = "task-1156092" [ 799.495568] env[62066]: _type = "Task" [ 799.495568] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.502688] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-vif-plugged-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.503464] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.503464] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.503464] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.503690] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] No waiting events found dispatching network-vif-plugged-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.503877] env[62066]: WARNING nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received unexpected event network-vif-plugged-4e3a7a96-3bfe-4fbe-9515-fd341f054700 for instance with vm_state building and task_state spawning. [ 799.504334] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.504408] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing instance network info cache due to event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.505474] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.507204] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.507422] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.507640] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.507948] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.508187] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.508430] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.508739] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.508999] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.509655] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.510031] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.510599] env[62066]: DEBUG nova.virt.hardware [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.512171] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b679ea1-3ca9-4dad-900d-e89d6a700fd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.519234] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.526176] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edeeb15-c993-4f18-ac19-b73e6aee25f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.545884] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-248d2815-9caf-4892-bd93-83a1e5d50633 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.579989] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Successfully created port: 2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.596891] env[62066]: DEBUG nova.network.neutron [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.681750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.297s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.689944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.714s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.689944] env[62066]: INFO nova.compute.claims [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.726394] env[62066]: INFO nova.scheduler.client.report [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Deleted allocations for instance 1c694323-616c-43b1-bd49-bba707e0788f [ 799.849930] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e0d77f-588a-9bd3-40eb-3f056bfe0a72, 'name': SearchDatastore_Task, 'duration_secs': 0.019524} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.850822] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7193413-935a-41db-868a-e38de3fdaed6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.857346] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 799.857346] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527a3de1-e9b1-ab97-e46d-edab485e958f" [ 799.857346] env[62066]: _type = "Task" [ 799.857346] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.866655] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a3de1-e9b1-ab97-e46d-edab485e958f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.884059] env[62066]: DEBUG oslo_vmware.api [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156090, 'name': PowerOnVM_Task, 'duration_secs': 0.751117} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.884385] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.884630] env[62066]: INFO nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Took 7.91 seconds to spawn the instance on the hypervisor. [ 799.886131] env[62066]: DEBUG nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.886131] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b9d0c6-aebf-4902-a814-7f35ec4cd526 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.006762] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156092, 'name': PowerOffVM_Task, 'duration_secs': 0.133826} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.007311] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 800.007447] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 800.007760] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bc65c89-bab7-46ed-8250-b7a5702fc807 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.035929] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 800.035929] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 800.035929] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Deleting the datastore file [datastore1] d9163e64-309b-4381-8819-15757f83ac2e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 800.035929] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c73acaf-f06a-4766-93a7-1788bc67c6aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.042483] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for the task: (returnval){ [ 800.042483] env[62066]: value = "task-1156094" [ 800.042483] env[62066]: _type = "Task" [ 800.042483] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.051070] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.079719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.080116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.081034] env[62066]: DEBUG nova.network.neutron [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.101100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.101577] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Instance network_info: |[{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.102160] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.102547] env[62066]: DEBUG nova.network.neutron [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.104418] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:6c:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271a82f1-1d09-4ad3-9c15-07269bad114c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e3a7a96-3bfe-4fbe-9515-fd341f054700', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.117703] env[62066]: DEBUG oslo.service.loopingcall [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.120331] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.120761] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8a6158d-46e0-4305-80b0-5ee7497f29a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.147388] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.147388] env[62066]: value = "task-1156095" [ 800.147388] env[62066]: _type = "Task" [ 800.147388] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.159523] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156095, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.238870] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0fd26441-5ba9-496b-8103-250bedcebaba tempest-ServerAddressesNegativeTestJSON-2027310797 tempest-ServerAddressesNegativeTestJSON-2027310797-project-member] Lock "1c694323-616c-43b1-bd49-bba707e0788f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.879s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.343870] env[62066]: DEBUG nova.network.neutron [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.374067] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a3de1-e9b1-ab97-e46d-edab485e958f, 'name': SearchDatastore_Task, 'duration_secs': 0.013284} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.374492] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.375984] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4723be94-3479-4e66-8088-914824c0e669/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 800.376790] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8071f0d4-7ca0-41bf-ac60-6264e469307a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.387096] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 800.387096] env[62066]: value = "task-1156096" [ 800.387096] env[62066]: _type = "Task" [ 800.387096] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.404060] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156096, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.407261] env[62066]: INFO nova.compute.manager [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Took 36.50 seconds to build instance. [ 800.554780] env[62066]: DEBUG oslo_vmware.api [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Task: {'id': task-1156094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162999} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.555186] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 800.555396] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 800.555877] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 800.555961] env[62066]: INFO nova.compute.manager [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Took 1.09 seconds to destroy the instance on the hypervisor. [ 800.556254] env[62066]: DEBUG oslo.service.loopingcall [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.557165] env[62066]: DEBUG nova.compute.manager [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.557306] env[62066]: DEBUG nova.network.neutron [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 800.576241] env[62066]: DEBUG nova.network.neutron [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.661605] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156095, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.848418] env[62066]: INFO nova.compute.manager [-] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Took 1.58 seconds to deallocate network for instance. [ 800.906458] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156096, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.911650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73aaea4c-0af5-47e8-9845-34ba2b228766 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.095s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.043987] env[62066]: INFO nova.network.neutron [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Port 8a5014e0-3b5b-46a1-b1b6-b98e168c945f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 801.044626] env[62066]: DEBUG nova.network.neutron [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.064557] env[62066]: DEBUG nova.network.neutron [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updated VIF entry in instance network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.065042] env[62066]: DEBUG nova.network.neutron [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.079809] env[62066]: DEBUG nova.network.neutron [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.107367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.107710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.107953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.108715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.108715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.110874] env[62066]: INFO nova.compute.manager [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Terminating instance [ 801.113902] env[62066]: DEBUG nova.compute.manager [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 801.113902] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.114927] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba04b601-e6f1-43e7-8cea-f594d820b39c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.126762] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 801.127156] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be0fe863-946a-4bb0-b4d0-22627a9fcbe2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.133855] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 801.133855] env[62066]: value = "task-1156097" [ 801.133855] env[62066]: _type = "Task" [ 801.133855] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.147439] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156097, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.158914] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156095, 'name': CreateVM_Task, 'duration_secs': 0.611838} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.162744] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.164364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.164572] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.164902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.166074] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8acf5114-8455-42bb-80f8-743884a66b34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.171152] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 801.171152] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52796939-5a5b-6489-1e13-16acb72850a4" [ 801.171152] env[62066]: _type = "Task" [ 801.171152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.187947] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52796939-5a5b-6489-1e13-16acb72850a4, 'name': SearchDatastore_Task, 'duration_secs': 0.013657} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.188382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.188625] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.188868] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.189034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.189474] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.189828] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fdbe4bf-2bc7-4ad3-82a5-f0d2cd34ac2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.199552] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.199767] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.200620] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcfc9f45-167a-4065-8bdb-00e351e3ba10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.207777] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 801.207777] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528f3c69-a599-edd2-d0c7-9d0892a411a0" [ 801.207777] env[62066]: _type = "Task" [ 801.207777] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.220562] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528f3c69-a599-edd2-d0c7-9d0892a411a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.250870] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff3daca-fa21-4ea2-a541-d533d798d409 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.261321] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ec2da3-bc73-42d3-ae63-05b1432a2cd2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.298916] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d937c2df-c042-4f1c-bf1a-f16a7b5db823 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.306866] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19da9f6f-e664-496e-8c97-d8f74ed7110d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.323016] env[62066]: DEBUG nova.compute.provider_tree [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.357382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.399294] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156096, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635944} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.399738] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4723be94-3479-4e66-8088-914824c0e669/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 801.401864] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d261590-420b-453e-8563-16dfc55d9c7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.436850] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 4723be94-3479-4e66-8088-914824c0e669/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.437878] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b52b1dc-fac9-4f1d-986a-4c5e917436af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.464545] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 801.464545] env[62066]: value = "task-1156098" [ 801.464545] env[62066]: _type = "Task" [ 801.464545] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.478396] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156098, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.547374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.567421] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.567902] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-vif-deleted-8a5014e0-3b5b-46a1-b1b6-b98e168c945f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.568092] env[62066]: INFO nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Neutron deleted interface 8a5014e0-3b5b-46a1-b1b6-b98e168c945f; detaching it from the instance and deleting it from the info cache [ 801.568438] env[62066]: DEBUG nova.network.neutron [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [{"id": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "address": "fa:16:3e:07:01:8b", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1436b8c7-b0", "ovs_interfaceid": "1436b8c7-b051-48ad-962f-30ccdc95b36b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.581545] env[62066]: INFO nova.compute.manager [-] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Took 1.02 seconds to deallocate network for instance. [ 801.645232] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156097, 'name': PowerOffVM_Task, 'duration_secs': 0.277288} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.645628] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.645885] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.646499] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-974ca121-50be-47b9-b932-5f37d348fa2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.719325] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528f3c69-a599-edd2-d0c7-9d0892a411a0, 'name': SearchDatastore_Task, 'duration_secs': 0.012893} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.720649] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d179ac04-ad97-4e2a-9c65-d1ffa3d06c8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.734095] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 801.734095] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528292ff-cce0-05b9-207b-4371c49a2502" [ 801.734095] env[62066]: _type = "Task" [ 801.734095] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.743944] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528292ff-cce0-05b9-207b-4371c49a2502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.826052] env[62066]: DEBUG nova.scheduler.client.report [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.977388] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156098, 'name': ReconfigVM_Task, 'duration_secs': 0.421146} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.977893] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 4723be94-3479-4e66-8088-914824c0e669/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.980560] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e55af1-e8bc-4068-a246-9cee38ffab01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.013220] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec31caf1-9bcc-40e2-abce-107206da72f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.032137] env[62066]: DEBUG nova.compute.manager [req-91fab8e2-9567-4365-9c7f-9aa97ce3e25e req-6fd3f001-b41d-4740-9cc5-c4fc47193d5a service nova] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Received event network-vif-deleted-9a995991-9d98-4fcc-a69b-cd7d9209d89f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.038492] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 802.038492] env[62066]: value = "task-1156100" [ 802.038492] env[62066]: _type = "Task" [ 802.038492] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.051502] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156100, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.054158] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d300c144-be28-444f-a1b4-dd63e2c7f9bc tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-26d87a85-0aa3-49b9-97ca-1b7fedbebb14-8a5014e0-3b5b-46a1-b1b6-b98e168c945f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.091s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.071750] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Acquiring lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.092723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.246433] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528292ff-cce0-05b9-207b-4371c49a2502, 'name': SearchDatastore_Task, 'duration_secs': 0.01445} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.246726] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.246990] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/831c0bed-8a41-4672-aa08-d60b64f365b8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.247338] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26e4429e-8e5b-4e81-b994-eb9c74d90861 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.255214] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 802.255214] env[62066]: value = "task-1156101" [ 802.255214] env[62066]: _type = "Task" [ 802.255214] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.265109] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156101, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.332065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.644s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.332743] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.335872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.113s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.339030] env[62066]: DEBUG nova.objects.instance [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lazy-loading 'resources' on Instance uuid 3d001331-7a41-4385-a622-54ce481509a8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.552873] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156100, 'name': ReconfigVM_Task, 'duration_secs': 0.2336} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.553362] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.553736] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7bf0c37-e3b4-4f06-a281-2239b74a8b86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.563856] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 802.563856] env[62066]: value = "task-1156102" [ 802.563856] env[62066]: _type = "Task" [ 802.563856] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.575787] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156102, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.663253] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Successfully updated port: b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.771148] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156101, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.839368] env[62066]: DEBUG nova.compute.utils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 802.845055] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.845326] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.030205] env[62066]: DEBUG nova.policy [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e276807bc06444eae42968bd7528926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a8e27dbc37043b3b36f8492c47d2286', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.046987] env[62066]: DEBUG nova.compute.manager [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-vif-plugged-b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.046987] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] Acquiring lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.047450] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.047807] env[62066]: DEBUG oslo_concurrency.lockutils [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.048143] env[62066]: DEBUG nova.compute.manager [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] No waiting events found dispatching network-vif-plugged-b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.048486] env[62066]: WARNING nova.compute.manager [req-cd166c03-be8c-4446-8dc6-316894879f83 req-66b00e24-7cee-4db1-92de-ef53d987a430 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received unexpected event network-vif-plugged-b2434f55-6f25-4684-82d8-eb2ac7745888 for instance with vm_state building and task_state spawning. [ 803.086105] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156102, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.269992] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156101, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.273827] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/831c0bed-8a41-4672-aa08-d60b64f365b8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.273827] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.273827] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b801941-c13c-4486-9cf1-d62b67587706 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.278448] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 803.278448] env[62066]: value = "task-1156103" [ 803.278448] env[62066]: _type = "Task" [ 803.278448] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.291304] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156103, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.349778] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.452623] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbfa9f4-4d32-4868-96b3-641de8f42d60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.460997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4ef0d0-2f08-4761-84fe-805a38e57dbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.495770] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb7ad7a-2a59-4714-85f0-9d93f9465c42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.504068] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8645c89d-7638-43a8-a510-5a255fbeb5b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.518956] env[62066]: DEBUG nova.compute.provider_tree [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.575906] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156102, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.794642] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156103, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069139} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.795735] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.796673] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b064da-52b5-4d18-94dc-7449aa2abf23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.801624] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.801827] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.802034] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleting the datastore file [datastore1] 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.802294] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9bd8fce-7013-4491-81d9-0855c830a73c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.825951] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/831c0bed-8a41-4672-aa08-d60b64f365b8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.826786] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-409d792a-e2f8-4dee-b804-558756479f6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.847825] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 803.847825] env[62066]: value = "task-1156104" [ 803.847825] env[62066]: _type = "Task" [ 803.847825] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.853726] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 803.853726] env[62066]: value = "task-1156105" [ 803.853726] env[62066]: _type = "Task" [ 803.853726] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.864642] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.871892] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.024369] env[62066]: DEBUG nova.scheduler.client.report [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.076847] env[62066]: DEBUG oslo_vmware.api [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156102, 'name': PowerOnVM_Task, 'duration_secs': 1.468841} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.077152] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.082160] env[62066]: DEBUG nova.compute.manager [None req-b31b2bff-dde2-4cb5-b1d3-4b1731fb6962 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.082160] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c6cbd1-fcac-4df4-b75e-01619c612dc7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.270033] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Successfully created port: 6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.361764] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.378463] env[62066]: DEBUG oslo_vmware.api [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.509585} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.378591] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.379642] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.380395] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.384021] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.384021] env[62066]: INFO nova.compute.manager [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Took 3.27 seconds to destroy the instance on the hypervisor. [ 804.384021] env[62066]: DEBUG oslo.service.loopingcall [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.384021] env[62066]: DEBUG nova.compute.manager [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.384021] env[62066]: DEBUG nova.network.neutron [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 804.529633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.532040] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.636s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.534291] env[62066]: INFO nova.compute.claims [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.559329] env[62066]: INFO nova.scheduler.client.report [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Deleted allocations for instance 3d001331-7a41-4385-a622-54ce481509a8 [ 804.867673] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156105, 'name': ReconfigVM_Task, 'duration_secs': 0.606785} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.868060] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/831c0bed-8a41-4672-aa08-d60b64f365b8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.869309] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5091e0d-8dc0-44c0-958f-4aaa9e690d57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.876132] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 804.876132] env[62066]: value = "task-1156106" [ 804.876132] env[62066]: _type = "Task" [ 804.876132] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.890677] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156106, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.074965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6381b5e8-f76e-4010-bc8a-e32190216dd2 tempest-ImagesNegativeTestJSON-1404790472 tempest-ImagesNegativeTestJSON-1404790472-project-member] Lock "3d001331-7a41-4385-a622-54ce481509a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.228s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.382109] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Successfully updated port: 2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.395831] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156106, 'name': Rename_Task, 'duration_secs': 0.186105} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.397228] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.397228] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15ad00d4-47e2-4460-a6ab-c6597c152769 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.405334] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 805.405334] env[62066]: value = "task-1156107" [ 805.405334] env[62066]: _type = "Task" [ 805.405334] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.416564] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.851200] env[62066]: DEBUG nova.compute.manager [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-changed-b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.851486] env[62066]: DEBUG nova.compute.manager [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Refreshing instance network info cache due to event network-changed-b2434f55-6f25-4684-82d8-eb2ac7745888. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.853163] env[62066]: DEBUG oslo_concurrency.lockutils [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] Acquiring lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.853163] env[62066]: DEBUG oslo_concurrency.lockutils [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] Acquired lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.853163] env[62066]: DEBUG nova.network.neutron [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Refreshing network info cache for port b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.890398] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.919672] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156107, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.967299] env[62066]: DEBUG nova.network.neutron [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.999617] env[62066]: DEBUG nova.compute.manager [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-vif-plugged-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.999922] env[62066]: DEBUG oslo_concurrency.lockutils [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] Acquiring lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.000230] env[62066]: DEBUG oslo_concurrency.lockutils [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.000608] env[62066]: DEBUG oslo_concurrency.lockutils [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.000849] env[62066]: DEBUG nova.compute.manager [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] No waiting events found dispatching network-vif-plugged-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.001051] env[62066]: WARNING nova.compute.manager [req-74063154-cbb9-469f-84c2-974264b04028 req-6f5772cc-caf4-4e13-8dba-578a617df24f service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received unexpected event network-vif-plugged-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 for instance with vm_state building and task_state spawning. [ 806.040823] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab14730-6002-4fa8-b5b0-0954ba4c2576 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.053486] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6fb022-fcd3-4a76-a8d2-8b018538df02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.102405] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdc0fd2-8bc0-4c8e-9f1c-8a3f042a8a35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.112318] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea6e6fd-d82f-4036-ae11-aa029e0d44b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.130333] env[62066]: DEBUG nova.compute.provider_tree [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.417884] env[62066]: DEBUG nova.network.neutron [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.426961] env[62066]: DEBUG oslo_vmware.api [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156107, 'name': PowerOnVM_Task, 'duration_secs': 0.713308} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.427263] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.427482] env[62066]: INFO nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Took 9.66 seconds to spawn the instance on the hypervisor. [ 806.427782] env[62066]: DEBUG nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.428710] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841f3de7-172c-4969-ac37-f82671b2fe12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.474888] env[62066]: INFO nova.compute.manager [-] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Took 2.09 seconds to deallocate network for instance. [ 806.508467] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.508467] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.508467] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.508648] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.508699] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.508841] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.509240] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.509473] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.509652] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.509813] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.509979] env[62066]: DEBUG nova.virt.hardware [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.512662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbec04b-9f7b-4253-9573-75dca2f6874a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.524850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ffb663-c8a8-401c-89f6-0745aaca3db6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.563863] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 806.567085] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577e0571-3ae1-4a20-bd5d-ed646ccb6310 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.571164] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 806.571340] env[62066]: ERROR oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk due to incomplete transfer. [ 806.571686] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-56324ffc-4fb5-41c7-8354-48690f81f90e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.583391] env[62066]: DEBUG nova.network.neutron [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.589122] env[62066]: DEBUG oslo_vmware.rw_handles [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524f9dd6-a0ce-c106-6239-231fde65bcf6/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 806.589341] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploaded image 59e9c507-e2ea-4178-b84f-f16bfba53d2a to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 806.594694] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 806.594694] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Successfully updated port: 6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.596020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f5509f1f-a148-4594-a6b4-6db715a19bbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.604074] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 806.604074] env[62066]: value = "task-1156108" [ 806.604074] env[62066]: _type = "Task" [ 806.604074] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.625019] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156108, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.633206] env[62066]: DEBUG nova.scheduler.client.report [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.959242] env[62066]: INFO nova.compute.manager [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Took 41.66 seconds to build instance. [ 806.986509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.090367] env[62066]: DEBUG oslo_concurrency.lockutils [req-32c06085-5ee7-405e-8e31-486926ed6a20 req-b2692716-db8f-4573-9873-58aa82b49bf3 service nova] Releasing lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.090753] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.090993] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.097665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.097665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.097665] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.116937] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156108, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.121999] env[62066]: INFO nova.compute.manager [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Rebuilding instance [ 807.143292] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.610s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.143572] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.155938] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.328s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.156324] env[62066]: DEBUG nova.objects.instance [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lazy-loading 'resources' on Instance uuid 92b150af-351a-43ab-952a-209225733ac8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.186444] env[62066]: DEBUG nova.compute.manager [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.187406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ac982a-9daf-40c6-b986-a17c94a780c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.464375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-431db74d-b64f-43f2-ad39-3eae9e8de922 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.170s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.567054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.567054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.614194] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156108, 'name': Destroy_Task, 'duration_secs': 0.715714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.614518] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroyed the VM [ 807.614828] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 807.615134] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-133d7a8b-1729-4c48-b106-47805f790d8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.622502] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 807.622502] env[62066]: value = "task-1156112" [ 807.622502] env[62066]: _type = "Task" [ 807.622502] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.631645] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156112, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.635071] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.650064] env[62066]: DEBUG nova.compute.utils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 807.651646] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.651803] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.662449] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.700266] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.700611] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9831d6e-e510-4e99-a0d5-26b5fcae76de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.710623] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 807.710623] env[62066]: value = "task-1156113" [ 807.710623] env[62066]: _type = "Task" [ 807.710623] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.721633] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156113, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.729834] env[62066]: DEBUG nova.policy [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53366fea3a6434fadadede81df16089', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33b756ddd30f4cb0b917fad171213266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.831307] env[62066]: DEBUG nova.network.neutron [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updating instance_info_cache with network_info: [{"id": "6383bfb0-3350-4704-bc27-2450027edc30", "address": "fa:16:3e:c4:b2:40", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6383bfb0-33", "ovs_interfaceid": "6383bfb0-3350-4704-bc27-2450027edc30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.069318] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 808.116587] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Successfully created port: 1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.124172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d0ec8e-c00d-4ccd-b001-eb4d323da6fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.145920] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d44e53-0cc1-463a-b796-96829613f017 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.149948] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156112, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.185102] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.192143] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba905db-ffa2-40a1-b5fe-a334ce1a5a0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.200424] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d59d35e-17b2-46b7-84f1-0f48a8723389 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.216494] env[62066]: DEBUG nova.compute.provider_tree [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.227837] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156113, 'name': PowerOffVM_Task, 'duration_secs': 0.219115} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.227948] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.228243] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.229292] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53c5921-2eeb-4288-b1a3-cd08c8bf3188 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.237909] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.238245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4b73bf3-07f3-4f53-a7d2-2828cb2f4f17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.267526] env[62066]: DEBUG nova.network.neutron [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updating instance_info_cache with network_info: [{"id": "b2434f55-6f25-4684-82d8-eb2ac7745888", "address": "fa:16:3e:f6:60:c8", "network": {"id": "0d685cc8-4498-4e49-ac29-55b24ef21e9c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-378550760", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2434f55-6f", "ovs_interfaceid": "b2434f55-6f25-4684-82d8-eb2ac7745888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "address": "fa:16:3e:3d:b8:dd", "network": {"id": "d3298ad1-f702-4e3e-b2e4-e2a40d7bbb59", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342438120", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2aee5757-0a", "ovs_interfaceid": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.300664] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.300664] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.300664] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore2] b46b7795-b7f7-4c55-bc72-184bb7bb8842 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.300899] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fdef762-5c89-44e2-96a5-28962d915e87 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.308033] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 808.308033] env[62066]: value = "task-1156115" [ 808.308033] env[62066]: _type = "Task" [ 808.308033] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.316936] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.334092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.334447] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Instance network_info: |[{"id": "6383bfb0-3350-4704-bc27-2450027edc30", "address": "fa:16:3e:c4:b2:40", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6383bfb0-33", "ovs_interfaceid": "6383bfb0-3350-4704-bc27-2450027edc30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.334899] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:b2:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6383bfb0-3350-4704-bc27-2450027edc30', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.343050] env[62066]: DEBUG oslo.service.loopingcall [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.343346] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.343603] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74c6fc47-9e5f-465b-b908-1a59bc978ea1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.363837] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.363837] env[62066]: value = "task-1156116" [ 808.363837] env[62066]: _type = "Task" [ 808.363837] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.372031] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156116, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.529343] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Received event network-vif-plugged-6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.529593] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Acquiring lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.529941] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.530109] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.530293] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] No waiting events found dispatching network-vif-plugged-6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.530651] env[62066]: WARNING nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Received unexpected event network-vif-plugged-6383bfb0-3350-4704-bc27-2450027edc30 for instance with vm_state building and task_state spawning. [ 808.530881] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Received event network-changed-6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.531064] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Refreshing instance network info cache due to event network-changed-6383bfb0-3350-4704-bc27-2450027edc30. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.531323] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Acquiring lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.531561] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Acquired lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.532133] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Refreshing network info cache for port 6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.596176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.618040] env[62066]: DEBUG nova.compute.manager [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-changed-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.618040] env[62066]: DEBUG nova.compute.manager [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Refreshing instance network info cache due to event network-changed-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.618040] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] Acquiring lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.647702] env[62066]: DEBUG oslo_vmware.api [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156112, 'name': RemoveSnapshot_Task, 'duration_secs': 0.80887} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.648140] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 808.648314] env[62066]: INFO nova.compute.manager [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 15.70 seconds to snapshot the instance on the hypervisor. [ 808.727021] env[62066]: DEBUG nova.scheduler.client.report [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.774029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.774258] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance network_info: |[{"id": "b2434f55-6f25-4684-82d8-eb2ac7745888", "address": "fa:16:3e:f6:60:c8", "network": {"id": "0d685cc8-4498-4e49-ac29-55b24ef21e9c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-378550760", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2434f55-6f", "ovs_interfaceid": "b2434f55-6f25-4684-82d8-eb2ac7745888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "address": "fa:16:3e:3d:b8:dd", "network": {"id": "d3298ad1-f702-4e3e-b2e4-e2a40d7bbb59", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342438120", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2aee5757-0a", "ovs_interfaceid": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.774618] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] Acquired lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.774814] env[62066]: DEBUG nova.network.neutron [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Refreshing network info cache for port 2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.776280] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:60:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2434f55-6f25-4684-82d8-eb2ac7745888', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:b8:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.786392] env[62066]: DEBUG oslo.service.loopingcall [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.789021] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.789572] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3184e2a-bd09-4561-be13-900804bfe605 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.819758] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201487} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.821388] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.821675] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.821875] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.824496] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.824496] env[62066]: value = "task-1156117" [ 808.824496] env[62066]: _type = "Task" [ 808.824496] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.833636] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156117, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.856246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.856482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.874746] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156116, 'name': CreateVM_Task, 'duration_secs': 0.372009} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.874930] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.875641] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.875812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.876244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.876718] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-084213a5-5dd5-401d-9eda-6107daab473a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.881413] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 808.881413] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5209f892-87c7-cf46-aedc-ab976d54b7d1" [ 808.881413] env[62066]: _type = "Task" [ 808.881413] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.892544] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5209f892-87c7-cf46-aedc-ab976d54b7d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.197662] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.233953] env[62066]: DEBUG nova.compute.manager [None req-2dcad4ea-0839-4c2b-ba57-3dec7e8a47ce tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Found 2 images (rotation: 2) {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 809.236261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.080s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.240805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.399s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.240805] env[62066]: DEBUG nova.objects.instance [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lazy-loading 'resources' on Instance uuid 248f5281-eb46-4be1-8642-28813c4b2622 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.253997] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.254765] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.255345] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.255345] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.255898] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.256040] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.256406] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.256731] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.257068] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.257374] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.257845] env[62066]: DEBUG nova.virt.hardware [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.259207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6243c19-b119-4e19-b5d8-fbdd3d1c2e09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.265300] env[62066]: INFO nova.scheduler.client.report [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted allocations for instance 92b150af-351a-43ab-952a-209225733ac8 [ 809.270863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6b716a-0cef-4840-8161-8ccbf4903fbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.279615] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updated VIF entry in instance network info cache for port 6383bfb0-3350-4704-bc27-2450027edc30. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.279966] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updating instance_info_cache with network_info: [{"id": "6383bfb0-3350-4704-bc27-2450027edc30", "address": "fa:16:3e:c4:b2:40", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6383bfb0-33", "ovs_interfaceid": "6383bfb0-3350-4704-bc27-2450027edc30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.282216] env[62066]: DEBUG nova.network.neutron [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updated VIF entry in instance network info cache for port 2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.282683] env[62066]: DEBUG nova.network.neutron [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updating instance_info_cache with network_info: [{"id": "b2434f55-6f25-4684-82d8-eb2ac7745888", "address": "fa:16:3e:f6:60:c8", "network": {"id": "0d685cc8-4498-4e49-ac29-55b24ef21e9c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-378550760", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.120", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2434f55-6f", "ovs_interfaceid": "b2434f55-6f25-4684-82d8-eb2ac7745888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "address": "fa:16:3e:3d:b8:dd", "network": {"id": "d3298ad1-f702-4e3e-b2e4-e2a40d7bbb59", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342438120", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2aee5757-0a", "ovs_interfaceid": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.340774] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156117, 'name': CreateVM_Task, 'duration_secs': 0.415428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.342245] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.345021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.345021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.345021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.345021] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e2dd1e-95f6-4515-b7f7-27ca06cced75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.353153] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 809.353153] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cc76e7-a6fa-9975-0868-c7f9c8b717e2" [ 809.353153] env[62066]: _type = "Task" [ 809.353153] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.361573] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.380302] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cc76e7-a6fa-9975-0868-c7f9c8b717e2, 'name': SearchDatastore_Task, 'duration_secs': 0.014493} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.381948] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.382604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.383214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.383706] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.384213] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.385528] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ea09a2f-45bc-46fa-8325-6aad59026cd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.405054] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5209f892-87c7-cf46-aedc-ab976d54b7d1, 'name': SearchDatastore_Task, 'duration_secs': 0.016917} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.405054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.405338] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.405667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.405829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.406102] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.408461] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e15b889-7c47-45ae-9686-ec120244e394 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.411352] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.411352] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.412509] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54dc08b4-e344-4a9a-b2d4-c0fb00676f05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.419977] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 809.419977] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ae46e7-04ba-ef1e-59eb-67dbefab36f7" [ 809.419977] env[62066]: _type = "Task" [ 809.419977] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.428363] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.428528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.429440] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e3f3a12-f40f-4d98-96dd-d53f03049f1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.438881] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ae46e7-04ba-ef1e-59eb-67dbefab36f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.447123] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 809.447123] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c1f522-e98c-d139-7ffd-51e0b9613705" [ 809.447123] env[62066]: _type = "Task" [ 809.447123] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.455516] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c1f522-e98c-d139-7ffd-51e0b9613705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.548237] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 809.549244] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f467d4-0652-43da-b756-8fd56d0b186f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.557522] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 809.557664] env[62066]: ERROR oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk due to incomplete transfer. [ 809.557899] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-452c47d4-00c8-4905-9687-6edcf67b4259 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.564375] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525827f4-aee4-89ab-38a7-28c367e92e01/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 809.564582] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Uploaded image ca07540e-1efc-45fb-a554-db31a70bc000 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 809.566258] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 809.566523] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bd1b0bc2-3669-4002-8231-ce9e1b255d3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.573391] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 809.573391] env[62066]: value = "task-1156118" [ 809.573391] env[62066]: _type = "Task" [ 809.573391] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.581638] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156118, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.691578] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Successfully updated port: 1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.781885] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Releasing lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.782139] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.782274] env[62066]: DEBUG nova.compute.manager [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing instance network info cache due to event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.782566] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.784089] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.784089] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.788939] env[62066]: DEBUG oslo_concurrency.lockutils [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] Releasing lock "refresh_cache-f196adc6-86bb-4556-bf3d-c52b9108c000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.788939] env[62066]: DEBUG nova.compute.manager [req-9fbf75b9-c08c-4ca5-8b5a-80d4533466d7 req-7efefe2f-9145-4135-8c37-ec05dbd08784 service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Received event network-vif-deleted-1436b8c7-b051-48ad-962f-30ccdc95b36b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.789614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9ef69120-47d3-45ac-8fb1-572ece785663 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "92b150af-351a-43ab-952a-209225733ac8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.446s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.860576] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.860834] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.861056] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.861294] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.861492] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.861643] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.861876] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.862055] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.862228] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.862394] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.862589] env[62066]: DEBUG nova.virt.hardware [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.863596] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97380fc2-383c-4cd6-b963-02eb8e0bfa83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.878988] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b84880f-625b-49d6-a843-f3b544e9aacf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.888662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.898386] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:3f:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.907190] env[62066]: DEBUG oslo.service.loopingcall [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.910446] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.910897] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95a71275-1499-4731-abf8-36e5781cb1dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.940252] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ae46e7-04ba-ef1e-59eb-67dbefab36f7, 'name': SearchDatastore_Task, 'duration_secs': 0.013808} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.942827] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.942827] env[62066]: value = "task-1156119" [ 809.942827] env[62066]: _type = "Task" [ 809.942827] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.942827] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18821342-f21c-478c-b6f3-b80751651e8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.956379] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 809.956379] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524636ad-9795-5aa7-b9f6-64c7c98ac744" [ 809.956379] env[62066]: _type = "Task" [ 809.956379] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.962254] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156119, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.971713] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c1f522-e98c-d139-7ffd-51e0b9613705, 'name': SearchDatastore_Task, 'duration_secs': 0.021784} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.973764] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d6c9be3-035e-4de9-b98b-c4f08a3c38b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.979094] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524636ad-9795-5aa7-b9f6-64c7c98ac744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.982271] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 809.982271] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52783bfd-9fb4-48f0-b68b-1a8a367b8487" [ 809.982271] env[62066]: _type = "Task" [ 809.982271] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.993447] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52783bfd-9fb4-48f0-b68b-1a8a367b8487, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.090830] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156118, 'name': Destroy_Task, 'duration_secs': 0.455205} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.091132] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Destroyed the VM [ 810.091375] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 810.091675] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-99c0f836-d102-41c8-8bbe-7ba6b6d613f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.100278] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 810.100278] env[62066]: value = "task-1156121" [ 810.100278] env[62066]: _type = "Task" [ 810.100278] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.109097] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156121, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.189482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.189654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.189832] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.250686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419f22d4-32c0-4b96-8739-0e47aa942834 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.260021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eddd5ab-4bcd-410c-bcc8-b66d6648e65b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.294573] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1edf45e-87d9-4361-9b33-b161562841ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.303315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e176a813-055b-4da1-8ae7-9989101673f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.318381] env[62066]: DEBUG nova.compute.provider_tree [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.470564] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156119, 'name': CreateVM_Task, 'duration_secs': 0.487396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.475622] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.475622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.475622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.475622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.475903] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e871f829-73d1-48e2-b08a-8133edbae2d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.481546] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524636ad-9795-5aa7-b9f6-64c7c98ac744, 'name': SearchDatastore_Task, 'duration_secs': 0.028865} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.482818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.482818] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f196adc6-86bb-4556-bf3d-c52b9108c000/f196adc6-86bb-4556-bf3d-c52b9108c000.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.482818] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-365d9034-c525-4f99-896e-ec7595d0c0c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.486421] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 810.486421] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd1935-0035-bde3-5d78-9939a1849d64" [ 810.486421] env[62066]: _type = "Task" [ 810.486421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.494487] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 810.494487] env[62066]: value = "task-1156122" [ 810.494487] env[62066]: _type = "Task" [ 810.494487] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.498496] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52783bfd-9fb4-48f0-b68b-1a8a367b8487, 'name': SearchDatastore_Task, 'duration_secs': 0.021801} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.504884] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.504884] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/f3ef43e8-6092-44ac-8990-979810a9748f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.505213] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd1935-0035-bde3-5d78-9939a1849d64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.508150] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb250359-b564-46f0-a162-60b50b2c0dbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.517359] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156122, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.518061] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 810.518061] env[62066]: value = "task-1156123" [ 810.518061] env[62066]: _type = "Task" [ 810.518061] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.525057] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.551754] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updated VIF entry in instance network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.553023] env[62066]: DEBUG nova.network.neutron [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.610555] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156121, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.729872] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.823952] env[62066]: DEBUG nova.scheduler.client.report [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.905654] env[62066]: DEBUG nova.network.neutron [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Updating instance_info_cache with network_info: [{"id": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "address": "fa:16:3e:70:35:49", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ab1ec14-b8", "ovs_interfaceid": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.008805] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd1935-0035-bde3-5d78-9939a1849d64, 'name': SearchDatastore_Task, 'duration_secs': 0.029427} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.009940] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.010788] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.012964] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.013302] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.013564] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.013999] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e34e6e66-d444-4e1c-8098-804aacc2fd35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.022385] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156122, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.039693] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156123, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.041385] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.041604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.042424] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d7e5983-b934-4ae4-919d-a2c4d716ad55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.048887] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 811.048887] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f50c67-00dd-c425-ee93-cc12d48639fb" [ 811.048887] env[62066]: _type = "Task" [ 811.048887] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.062446] env[62066]: DEBUG oslo_concurrency.lockutils [req-2be7637f-b755-4ad5-b5c9-923fdecbbeb4 req-b606f58c-52d7-4233-a861-5a3d7f9952f4 service nova] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.063267] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f50c67-00dd-c425-ee93-cc12d48639fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.113548] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156121, 'name': RemoveSnapshot_Task, 'duration_secs': 0.739291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.113813] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 811.114188] env[62066]: DEBUG nova.compute.manager [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.115206] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042674d9-6cd8-4ac4-875d-08e33d83143b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.332679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.092s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.338505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.489s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.338740] env[62066]: DEBUG nova.objects.instance [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lazy-loading 'resources' on Instance uuid c6fad95f-9167-4ad3-9c1c-16425fa1d36c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.359470] env[62066]: INFO nova.scheduler.client.report [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted allocations for instance 248f5281-eb46-4be1-8642-28813c4b2622 [ 811.375905] env[62066]: DEBUG nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.377759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652a6042-3648-4f32-942c-286bea4d2f36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.411662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.412038] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Instance network_info: |[{"id": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "address": "fa:16:3e:70:35:49", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ab1ec14-b8", "ovs_interfaceid": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.412489] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:35:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ab1ec14-b82d-41c3-880e-3aed3469b3a3', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.421507] env[62066]: DEBUG oslo.service.loopingcall [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.422831] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.423538] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcc46779-1988-410a-b5e7-8ffd4418fd75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.444749] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.444749] env[62066]: value = "task-1156124" [ 811.444749] env[62066]: _type = "Task" [ 811.444749] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.454455] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156124, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.509198] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156122, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.789704} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.509611] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f196adc6-86bb-4556-bf3d-c52b9108c000/f196adc6-86bb-4556-bf3d-c52b9108c000.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.509888] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.510183] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc16a5d1-2b4a-4fa6-9d2d-258ab6bfcd48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.516229] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 811.516229] env[62066]: value = "task-1156125" [ 811.516229] env[62066]: _type = "Task" [ 811.516229] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.527145] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156125, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.536227] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156123, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.536499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/f3ef43e8-6092-44ac-8990-979810a9748f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.536721] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.536983] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d9392cc-58e3-4b72-b02f-5688a8edfa05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.542674] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 811.542674] env[62066]: value = "task-1156126" [ 811.542674] env[62066]: _type = "Task" [ 811.542674] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.550828] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.558970] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f50c67-00dd-c425-ee93-cc12d48639fb, 'name': SearchDatastore_Task, 'duration_secs': 0.0688} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.559767] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddc52b2e-8855-413d-9b80-b90328e55d23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.564914] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 811.564914] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524a13eb-0fed-aa18-0310-6950f461c1b7" [ 811.564914] env[62066]: _type = "Task" [ 811.564914] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.573055] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524a13eb-0fed-aa18-0310-6950f461c1b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.633519] env[62066]: INFO nova.compute.manager [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Shelve offloading [ 811.634087] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 811.634342] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38719c18-7b49-4386-9e93-f9fe52daa8ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.641918] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 811.641918] env[62066]: value = "task-1156127" [ 811.641918] env[62066]: _type = "Task" [ 811.641918] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.650322] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.679954] env[62066]: DEBUG nova.compute.manager [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.680187] env[62066]: DEBUG nova.compute.manager [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing instance network info cache due to event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.680414] env[62066]: DEBUG oslo_concurrency.lockutils [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.680788] env[62066]: DEBUG oslo_concurrency.lockutils [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.680858] env[62066]: DEBUG nova.network.neutron [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.698918] env[62066]: DEBUG nova.compute.manager [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Received event network-vif-plugged-1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.699207] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Acquiring lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.699379] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.699558] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.699732] env[62066]: DEBUG nova.compute.manager [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] No waiting events found dispatching network-vif-plugged-1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.699894] env[62066]: WARNING nova.compute.manager [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Received unexpected event network-vif-plugged-1ab1ec14-b82d-41c3-880e-3aed3469b3a3 for instance with vm_state building and task_state spawning. [ 811.700070] env[62066]: DEBUG nova.compute.manager [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Received event network-changed-1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.700284] env[62066]: DEBUG nova.compute.manager [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Refreshing instance network info cache due to event network-changed-1ab1ec14-b82d-41c3-880e-3aed3469b3a3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.700414] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Acquiring lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.700580] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Acquired lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.700750] env[62066]: DEBUG nova.network.neutron [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Refreshing network info cache for port 1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.871284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4eaaeb46-523a-4c2e-afc1-c8a04b2a0734 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "248f5281-eb46-4be1-8642-28813c4b2622" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.220s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.890103] env[62066]: INFO nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] instance snapshotting [ 811.890693] env[62066]: DEBUG nova.objects.instance [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.955609] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156124, 'name': CreateVM_Task, 'duration_secs': 0.319866} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.955783] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.956532] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.956712] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.957069] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.957329] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e841cea-70b8-451e-bff8-337698300f01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.966029] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 811.966029] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52526937-afeb-8377-c7e6-93dd2d919c9e" [ 811.966029] env[62066]: _type = "Task" [ 811.966029] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.975092] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52526937-afeb-8377-c7e6-93dd2d919c9e, 'name': SearchDatastore_Task, 'duration_secs': 0.008456} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.975719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.975972] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.976226] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.030837] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156125, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079578} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.031145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.031934] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16545ab4-7c0c-499a-ace3-552213788156 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.059826] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] f196adc6-86bb-4556-bf3d-c52b9108c000/f196adc6-86bb-4556-bf3d-c52b9108c000.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.066299] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eedcf39f-2927-4ab6-b1c6-03274d6e7bb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.091969] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067582} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.096413] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.096770] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524a13eb-0fed-aa18-0310-6950f461c1b7, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.097051] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 812.097051] env[62066]: value = "task-1156129" [ 812.097051] env[62066]: _type = "Task" [ 812.097051] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.097747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edce8541-1857-495f-b6fd-70bd5bc17a77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.100916] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.100916] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.103179] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.103379] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.103594] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1bf24ae-28d6-4658-919d-c9c1f3e280a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.110206] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30d3dc65-e9db-48d6-b5c4-f7a684ad23d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.131867] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/f3ef43e8-6092-44ac-8990-979810a9748f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.140651] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd8ac548-a37c-4165-8619-c0550d960c7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.155859] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 812.155859] env[62066]: value = "task-1156130" [ 812.155859] env[62066]: _type = "Task" [ 812.155859] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.156162] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.160028] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.160028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.165597] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4f66388-ac39-4404-99de-5bd4b55d282f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.171675] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 812.171675] env[62066]: value = "task-1156131" [ 812.171675] env[62066]: _type = "Task" [ 812.171675] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.185647] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.188098] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 812.188098] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5228efde-279d-a1fb-2d8c-cfc4d6a75475" [ 812.188098] env[62066]: _type = "Task" [ 812.188098] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.189214] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 812.189214] env[62066]: DEBUG nova.compute.manager [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.190659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2ac91e-360a-4a3d-bf19-a5bae99777f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.202503] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.210239] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5228efde-279d-a1fb-2d8c-cfc4d6a75475, 'name': SearchDatastore_Task, 'duration_secs': 0.010574} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.210587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.210790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.210972] env[62066]: DEBUG nova.network.neutron [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.213429] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97afd65a-ac83-4975-9609-e3868d44d1f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.221876] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 812.221876] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ba9f37-31ce-b253-0b36-f89c200eb0de" [ 812.221876] env[62066]: _type = "Task" [ 812.221876] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.231210] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ba9f37-31ce-b253-0b36-f89c200eb0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.379686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbd5bf4-9ce3-4811-b14f-b7dc2d287b5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.391654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40931026-1696-4744-b567-8e5f8c068e2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.399092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719dab77-886b-4f76-bfa6-35a5c2adf7bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.443241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25486be-bae9-43d1-8855-299308b4cf18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.461914] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93a206b-083d-432b-9ef9-2a230845b7c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.468762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0550770f-812f-48d9-8cbe-35ba5b4c6ebf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.489168] env[62066]: DEBUG nova.compute.provider_tree [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.518393] env[62066]: DEBUG nova.network.neutron [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updated VIF entry in instance network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 812.518945] env[62066]: DEBUG nova.network.neutron [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.612411] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.669219] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474219} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.669650] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.669991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.670364] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1986965a-a4bf-4c55-811b-174b96cd05bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.683615] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.684970] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 812.684970] env[62066]: value = "task-1156132" [ 812.684970] env[62066]: _type = "Task" [ 812.684970] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.692904] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.733104] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ba9f37-31ce-b253-0b36-f89c200eb0de, 'name': SearchDatastore_Task, 'duration_secs': 0.01795} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.733480] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.733700] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 854a1ea4-515d-4a05-b179-22713b63f7c3/854a1ea4-515d-4a05-b179-22713b63f7c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.734026] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5c2e574-9360-4b48-b7bf-411cb793e303 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.740096] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 812.740096] env[62066]: value = "task-1156133" [ 812.740096] env[62066]: _type = "Task" [ 812.740096] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.749956] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.955172] env[62066]: DEBUG nova.network.neutron [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.980135] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 812.980610] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a7e23712-33f5-41b4-b6a2-3e7d1ba039f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.993802] env[62066]: DEBUG nova.scheduler.client.report [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.998994] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 812.998994] env[62066]: value = "task-1156134" [ 812.998994] env[62066]: _type = "Task" [ 812.998994] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.010529] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156134, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.026135] env[62066]: DEBUG oslo_concurrency.lockutils [req-041a9b13-3b79-4100-8f7e-a13a53c05ea1 req-48977b0b-3558-42c9-a9d9-dc1b205dc12c service nova] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.112899] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156129, 'name': ReconfigVM_Task, 'duration_secs': 0.711011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.113374] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Reconfigured VM instance instance-00000044 to attach disk [datastore2] f196adc6-86bb-4556-bf3d-c52b9108c000/f196adc6-86bb-4556-bf3d-c52b9108c000.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.114322] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-550d4df5-0b8f-4fb6-b1de-a0c46b913514 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.122706] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 813.122706] env[62066]: value = "task-1156135" [ 813.122706] env[62066]: _type = "Task" [ 813.122706] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.129029] env[62066]: DEBUG nova.network.neutron [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Updated VIF entry in instance network info cache for port 1ab1ec14-b82d-41c3-880e-3aed3469b3a3. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.129665] env[62066]: DEBUG nova.network.neutron [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Updating instance_info_cache with network_info: [{"id": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "address": "fa:16:3e:70:35:49", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ab1ec14-b8", "ovs_interfaceid": "1ab1ec14-b82d-41c3-880e-3aed3469b3a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.139045] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156135, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.185819] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156131, 'name': ReconfigVM_Task, 'duration_secs': 0.702925} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.186896] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfigured VM instance instance-00000045 to attach disk [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/f3ef43e8-6092-44ac-8990-979810a9748f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.196156] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e277ceb5-0d7b-4588-9a86-57d4237c2700 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.209067] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19124} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.209067] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.209067] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 813.209067] env[62066]: value = "task-1156136" [ 813.209067] env[62066]: _type = "Task" [ 813.209067] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.209067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfbc255-0b7d-497d-9605-59d68055b4bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.243297] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.243782] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156136, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.244112] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9321f8a1-b273-4e4a-9352-9773b3e6f44f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.268098] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4833} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.269412] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 854a1ea4-515d-4a05-b179-22713b63f7c3/854a1ea4-515d-4a05-b179-22713b63f7c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.269634] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.269940] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 813.269940] env[62066]: value = "task-1156137" [ 813.269940] env[62066]: _type = "Task" [ 813.269940] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.270142] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce14fcb0-662f-458f-8345-1c88b7998b8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.280296] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156137, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.281572] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 813.281572] env[62066]: value = "task-1156138" [ 813.281572] env[62066]: _type = "Task" [ 813.281572] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.289525] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156138, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.458324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.500800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.162s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.503545] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.154s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.503760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.510569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.328s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.512202] env[62066]: INFO nova.compute.claims [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.532120] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156134, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.543801] env[62066]: INFO nova.scheduler.client.report [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Deleted allocations for instance c6fad95f-9167-4ad3-9c1c-16425fa1d36c [ 813.574865] env[62066]: INFO nova.scheduler.client.report [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Deleted allocations for instance c0ac4362-766f-48ba-aeb2-7fd976c1f47f [ 813.633977] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156135, 'name': Rename_Task, 'duration_secs': 0.248858} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.634293] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.634538] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db8bbcd6-297d-433d-8d23-f34d2a71a937 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.641212] env[62066]: DEBUG oslo_concurrency.lockutils [req-407f2ee7-b78f-4561-a704-fa8b798dafb1 req-598f76dd-2c22-4957-807a-56dfa22d94ea service nova] Releasing lock "refresh_cache-854a1ea4-515d-4a05-b179-22713b63f7c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.641663] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 813.641663] env[62066]: value = "task-1156139" [ 813.641663] env[62066]: _type = "Task" [ 813.641663] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.653231] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156139, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.729085] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156136, 'name': Rename_Task, 'duration_secs': 0.28544} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.729430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.729687] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8636d27-8d5b-49a1-9901-ee6291b48921 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.736575] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 813.736575] env[62066]: value = "task-1156140" [ 813.736575] env[62066]: _type = "Task" [ 813.736575] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.744364] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.781834] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156137, 'name': ReconfigVM_Task, 'duration_secs': 0.278637} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.782211] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Reconfigured VM instance instance-00000042 to attach disk [datastore1] b46b7795-b7f7-4c55-bc72-184bb7bb8842/b46b7795-b7f7-4c55-bc72-184bb7bb8842.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.787476] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-513fc666-08dc-4f17-891a-5b196182f0f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.793416] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156138, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074622} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.794991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.795373] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 813.795373] env[62066]: value = "task-1156141" [ 813.795373] env[62066]: _type = "Task" [ 813.795373] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.796081] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e305e9-2a62-4d27-b071-8db5cc065939 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.812027] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156141, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.833433] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 854a1ea4-515d-4a05-b179-22713b63f7c3/854a1ea4-515d-4a05-b179-22713b63f7c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.833747] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-089576f6-2f14-49e9-b6ae-20159763cc9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.854213] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 813.854213] env[62066]: value = "task-1156142" [ 813.854213] env[62066]: _type = "Task" [ 813.854213] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.863373] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156142, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.013681] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156134, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.055733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc5fb12-eb8d-43ed-947e-0346ec252917 tempest-MultipleCreateTestJSON-1362050231 tempest-MultipleCreateTestJSON-1362050231-project-member] Lock "c6fad95f-9167-4ad3-9c1c-16425fa1d36c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.176s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.088384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22560994-a94c-4905-95c1-ce7724e7e424 tempest-MigrationsAdminTest-1258103067 tempest-MigrationsAdminTest-1258103067-project-member] Lock "c0ac4362-766f-48ba-aeb2-7fd976c1f47f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.769s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.157686] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156139, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.248032] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156140, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.311017] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156141, 'name': Rename_Task, 'duration_secs': 0.162063} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.311017] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.311017] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62a48270-f26d-47bd-91c4-d4919160285d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.317121] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 814.317121] env[62066]: value = "task-1156144" [ 814.317121] env[62066]: _type = "Task" [ 814.317121] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.328413] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.367140] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156142, 'name': ReconfigVM_Task, 'duration_secs': 0.297313} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.367140] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 854a1ea4-515d-4a05-b179-22713b63f7c3/854a1ea4-515d-4a05-b179-22713b63f7c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.367140] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb919777-1e2f-4ce3-b10f-b8242ddff158 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.373656] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 814.373656] env[62066]: value = "task-1156145" [ 814.373656] env[62066]: _type = "Task" [ 814.373656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.388238] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156145, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.517193] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156134, 'name': CreateSnapshot_Task, 'duration_secs': 1.081864} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.517761] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 814.520620] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c907314-b867-4bfe-9477-4cdf43a72df4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.544333] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 814.545254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2bfe624-6ba4-4e78-a5a4-c0b34b7f1288 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.553982] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 814.554601] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61bbcbeb-01a6-482e-84e8-97b76e255cfa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.638490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 814.638778] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 814.638932] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleting the datastore file [datastore1] a2251e56-2787-412d-89c9-eef111ee6d2b {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.639225] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-999e1735-e32c-400d-9edd-ba34ff1493c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.654751] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 814.654751] env[62066]: value = "task-1156147" [ 814.654751] env[62066]: _type = "Task" [ 814.654751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.662813] env[62066]: DEBUG oslo_vmware.api [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156139, 'name': PowerOnVM_Task, 'duration_secs': 0.941113} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.665839] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.666116] env[62066]: INFO nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Took 15.20 seconds to spawn the instance on the hypervisor. [ 814.666316] env[62066]: DEBUG nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.667362] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f6c9c3-c986-4e17-b759-e6b355893c81 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.673162] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.750761] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156140, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.810576] env[62066]: DEBUG nova.compute.manager [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.810576] env[62066]: DEBUG nova.compute.manager [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing instance network info cache due to event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.810576] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.810684] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.810866] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.833178] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156144, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.890816] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156145, 'name': Rename_Task, 'duration_secs': 0.212651} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.891144] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.891748] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58753619-768e-4aab-b012-b64108770272 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.898897] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 814.898897] env[62066]: value = "task-1156148" [ 814.898897] env[62066]: _type = "Task" [ 814.898897] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.913450] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.983300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcca5def-4394-4064-a562-a853a2d78560 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.993813] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba89fbc-b1bd-4c9f-800e-9af5ee5294db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.032179] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64059bf-4486-4bc2-b12d-c7d8287d2dde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.039487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae97c85-1a40-40bc-a568-4d2c9120ea9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.059627] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 815.060016] env[62066]: DEBUG nova.compute.provider_tree [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.061283] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c44cdc9b-8a58-4afb-8c04-531324a1344c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.069674] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 815.069674] env[62066]: value = "task-1156149" [ 815.069674] env[62066]: _type = "Task" [ 815.069674] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.080683] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156149, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.169639] env[62066]: DEBUG oslo_vmware.api [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266412} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.169639] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.169639] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 815.170135] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 815.210950] env[62066]: INFO nova.compute.manager [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Took 46.34 seconds to build instance. [ 815.220102] env[62066]: INFO nova.scheduler.client.report [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted allocations for instance a2251e56-2787-412d-89c9-eef111ee6d2b [ 815.255699] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156140, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.332610] env[62066]: DEBUG oslo_vmware.api [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156144, 'name': PowerOnVM_Task, 'duration_secs': 0.670544} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.333123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.333456] env[62066]: DEBUG nova.compute.manager [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.335943] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6124f325-6a54-4d3a-9725-acfd6664e8ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.412934] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156148, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.565961] env[62066]: DEBUG nova.scheduler.client.report [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.579907] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156149, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.722462] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73614f60-a4d8-460f-929a-f8cda300674f tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.462s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.724046] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.751684] env[62066]: DEBUG oslo_vmware.api [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156140, 'name': PowerOnVM_Task, 'duration_secs': 1.972941} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.752297] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.753634] env[62066]: INFO nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Took 11.39 seconds to spawn the instance on the hypervisor. [ 815.753966] env[62066]: DEBUG nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.755206] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618e1c16-a7c3-4ba2-be10-096a581a7a4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.860698] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.915178] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156148, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.071847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.072673] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.081145] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.248s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.100037] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156149, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.289025] env[62066]: INFO nova.compute.manager [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Took 46.33 seconds to build instance. [ 816.353296] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updated VIF entry in instance network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 816.353296] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.417784] env[62066]: DEBUG oslo_vmware.api [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156148, 'name': PowerOnVM_Task, 'duration_secs': 1.109336} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.419023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.419023] env[62066]: INFO nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Took 7.22 seconds to spawn the instance on the hypervisor. [ 816.419023] env[62066]: DEBUG nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.419497] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f392b9f-c581-4ad6-b23e-ec383b5f8f55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.590982] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156149, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.605885] env[62066]: DEBUG nova.compute.utils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 816.608849] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.608849] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.767334] env[62066]: DEBUG nova.policy [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8bcc97a71b4052b1ed4ba1b4194ff6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd61d04b8fe994dcfaa361bb87eb587d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.788628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2e8a55a-d580-4e0b-935e-36bbc3e466db tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.522s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.856570] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.856925] env[62066]: DEBUG nova.compute.manager [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.857171] env[62066]: DEBUG nova.compute.manager [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing instance network info cache due to event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.858391] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.858391] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.858391] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.960377] env[62066]: INFO nova.compute.manager [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Took 32.09 seconds to build instance. [ 817.090194] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156149, 'name': CloneVM_Task, 'duration_secs': 1.537886} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.090446] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Created linked-clone VM from snapshot [ 817.091338] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71490d8a-4e18-494e-b257-4202db5ef3ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.101373] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploading image 7edbe552-3036-4d5b-a9e5-62e23f253c7a {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 817.116077] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.149072] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 817.149072] env[62066]: value = "vm-251691" [ 817.149072] env[62066]: _type = "VirtualMachine" [ 817.149072] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 817.149716] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-94246664-8a97-4ea2-b6b9-c541de7e9975 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.154153] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.154385] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.154550] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2f186d0a-91a0-4dc2-83bd-511099445af7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.155048] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.155204] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d50f0505-66f4-412f-9744-25e0ea96277a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.155349] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a8dd7483-0588-4f60-9504-20de799e69f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.155470] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.155618] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.155739] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance de9eb96d-e71b-4878-83b8-f75966dc3f48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.155864] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.156024] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d9163e64-309b-4381-8819-15757f83ac2e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.156271] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.156338] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance e95ee795-603e-4cbf-bcd6-1ba54b62a281 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.156401] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 43d64f76-7f7a-4b95-b9df-c95218612998 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 817.156508] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 4723be94-3479-4e66-8088-914824c0e669 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.156640] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance b46b7795-b7f7-4c55-bc72-184bb7bb8842 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.156902] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 831c0bed-8a41-4672-aa08-d60b64f365b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.156902] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f196adc6-86bb-4556-bf3d-c52b9108c000 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.157015] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f3ef43e8-6092-44ac-8990-979810a9748f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.157142] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 854a1ea4-515d-4a05-b179-22713b63f7c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.157271] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 817.164611] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease: (returnval){ [ 817.164611] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d93701-4c9d-8ccd-5600-c1552457ef88" [ 817.164611] env[62066]: _type = "HttpNfcLease" [ 817.164611] env[62066]: } obtained for exporting VM: (result){ [ 817.164611] env[62066]: value = "vm-251691" [ 817.164611] env[62066]: _type = "VirtualMachine" [ 817.164611] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 817.164941] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the lease: (returnval){ [ 817.164941] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d93701-4c9d-8ccd-5600-c1552457ef88" [ 817.164941] env[62066]: _type = "HttpNfcLease" [ 817.164941] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 817.175914] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.175914] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d93701-4c9d-8ccd-5600-c1552457ef88" [ 817.175914] env[62066]: _type = "HttpNfcLease" [ 817.175914] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 817.439246] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Successfully created port: 03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.464658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b711efe2-87dd-4543-ba66-bf09bdbe84bf tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.602s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.534980] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.535254] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing instance network info cache due to event network-changed-4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.535468] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Acquiring lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.660399] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 68bf0e89-8236-44ef-b514-d25698eb23c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 817.674802] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.674802] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d93701-4c9d-8ccd-5600-c1552457ef88" [ 817.674802] env[62066]: _type = "HttpNfcLease" [ 817.674802] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 817.674802] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 817.674802] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d93701-4c9d-8ccd-5600-c1552457ef88" [ 817.674802] env[62066]: _type = "HttpNfcLease" [ 817.674802] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 817.675262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86910ab2-b4e6-4355-ad28-007a3fdea096 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.683873] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 817.684114] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 817.784242] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updated VIF entry in instance network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.784671] env[62066]: DEBUG nova.network.neutron [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.805043] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-46b9b0a8-9f83-4201-8971-203a01813ae8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.918408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "f196adc6-86bb-4556-bf3d-c52b9108c000" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.919252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.919252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.919252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.919474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.921399] env[62066]: INFO nova.compute.manager [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Terminating instance [ 817.924720] env[62066]: DEBUG nova.compute.manager [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.924917] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.925770] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86424da2-075f-4353-800d-b28fc84892bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.933245] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.933503] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b8604ce-0d67-4839-881c-eb5acc3c6cb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.939886] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 817.939886] env[62066]: value = "task-1156152" [ 817.939886] env[62066]: _type = "Task" [ 817.939886] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.950682] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.130344] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.164561] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3092f591-214f-40ba-a8e2-ccc6a0f007c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 818.164885] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 818.165746] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3392MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 818.171064] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.171064] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.171248] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.171485] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.171794] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.171794] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.172226] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.172422] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.172671] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.172901] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.173056] env[62066]: DEBUG nova.virt.hardware [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.174576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dd41b4-6c89-4c92-891f-4d8030a4bec6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.185613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630b0d01-5612-460d-aa5a-eb147791ecfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.289757] env[62066]: DEBUG oslo_concurrency.lockutils [req-6d1a4744-8f61-45dc-9030-532a6464de87 req-63565095-8d92-4677-b5bc-acf3c470ff67 service nova] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.290841] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Acquired lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.291125] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Refreshing network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.453612] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156152, 'name': PowerOffVM_Task, 'duration_secs': 0.233043} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.453940] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.454124] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.454438] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d008980e-dba4-4cf1-bc64-9a174b16542a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.467447] env[62066]: INFO nova.compute.manager [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Rescuing [ 818.467447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.467447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.467447] env[62066]: DEBUG nova.network.neutron [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.536828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "4723be94-3479-4e66-8088-914824c0e669" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.536828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.536828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "4723be94-3479-4e66-8088-914824c0e669-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.536828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.536828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.541014] env[62066]: INFO nova.compute.manager [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Terminating instance [ 818.547113] env[62066]: DEBUG nova.compute.manager [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.547113] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.547113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee92a5e0-4852-4224-b631-a7ce7fb44c8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.557362] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.558026] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4ce8da0-b194-497c-be2f-e27ab1e9c40b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.565843] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 818.565843] env[62066]: value = "task-1156155" [ 818.565843] env[62066]: _type = "Task" [ 818.565843] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.574699] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.579738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409889d6-9b5f-4cd1-9687-a202486b0ca6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.587278] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505f952c-8725-4b95-9297-61913cf46136 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.624911] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa46bc2-936e-4862-a358-d2a368f590e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.627879] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 818.628153] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 818.628380] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleting the datastore file [datastore2] f196adc6-86bb-4556-bf3d-c52b9108c000 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.628835] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df7543d2-53b8-4dc2-9478-7cad1513cd2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.636979] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e344b79-ca42-4770-bb0f-bd7ea60dc2b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.641651] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for the task: (returnval){ [ 818.641651] env[62066]: value = "task-1156156" [ 818.641651] env[62066]: _type = "Task" [ 818.641651] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.656044] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.663529] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156156, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.741805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.741805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.741805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.741805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.741805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.743914] env[62066]: INFO nova.compute.manager [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Terminating instance [ 818.746376] env[62066]: DEBUG nova.compute.manager [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.746376] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.747984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf971000-be68-43a6-b029-4c7b8fa8f2d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.756531] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.756927] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06dabbad-f571-460d-bb5f-187fd2156fd1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.765939] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 818.765939] env[62066]: value = "task-1156157" [ 818.765939] env[62066]: _type = "Task" [ 818.765939] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.776256] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.992021] env[62066]: DEBUG nova.compute.manager [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.992021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c07701e-3ce9-47fb-bad3-c43151b365e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.077079] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156155, 'name': PowerOffVM_Task, 'duration_secs': 0.263678} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.080415] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.080808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.083395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-605cf850-4494-4690-8f96-81ee9cbefa6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.154286] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.154858] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.156221] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Deleting the datastore file [datastore1] 4723be94-3479-4e66-8088-914824c0e669 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.156696] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updated VIF entry in instance network info cache for port 4d1d88f2-e881-41ed-8de8-7cf13954abf9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 819.156696] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [{"id": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "address": "fa:16:3e:4c:1e:fe", "network": {"id": "f24dd468-9459-40ee-b5d9-927160aa489c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-673767946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "720fe50954804162a95423afa176e526", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d1d88f2-e8", "ovs_interfaceid": "4d1d88f2-e881-41ed-8de8-7cf13954abf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.159165] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-819d59d7-50ba-4875-b5dc-a02173d1c0fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.169351] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.173178] env[62066]: DEBUG oslo_vmware.api [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Task: {'id': task-1156156, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.174655] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.174951] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.175263] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.175518] env[62066]: INFO nova.compute.manager [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Took 1.25 seconds to destroy the instance on the hypervisor. [ 819.175891] env[62066]: DEBUG oslo.service.loopingcall [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.179238] env[62066]: DEBUG nova.compute.manager [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.179238] env[62066]: DEBUG nova.network.neutron [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.182926] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for the task: (returnval){ [ 819.182926] env[62066]: value = "task-1156159" [ 819.182926] env[62066]: _type = "Task" [ 819.182926] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.192995] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.278856] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156157, 'name': PowerOffVM_Task, 'duration_secs': 0.303659} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.278856] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.278856] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.278856] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9357725-77ec-4573-b1ca-fa153b00f60a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.315522] env[62066]: DEBUG nova.network.neutron [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updating instance_info_cache with network_info: [{"id": "6383bfb0-3350-4704-bc27-2450027edc30", "address": "fa:16:3e:c4:b2:40", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6383bfb0-33", "ovs_interfaceid": "6383bfb0-3350-4704-bc27-2450027edc30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.346198] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.346556] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.346850] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore1] b46b7795-b7f7-4c55-bc72-184bb7bb8842 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.347253] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30d5d514-61c4-4db2-ae58-d02c02846727 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.355583] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 819.355583] env[62066]: value = "task-1156161" [ 819.355583] env[62066]: _type = "Task" [ 819.355583] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.364239] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.467917] env[62066]: DEBUG nova.compute.manager [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Received event network-vif-plugged-03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.468156] env[62066]: DEBUG oslo_concurrency.lockutils [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] Acquiring lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.468369] env[62066]: DEBUG oslo_concurrency.lockutils [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.468735] env[62066]: DEBUG oslo_concurrency.lockutils [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.469526] env[62066]: DEBUG nova.compute.manager [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] No waiting events found dispatching network-vif-plugged-03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.469720] env[62066]: WARNING nova.compute.manager [req-eeddb632-2f9d-4ed8-9cb9-d6f247677718 req-44dcf729-fe4a-4247-a025-bf2c2a7e14e7 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Received unexpected event network-vif-plugged-03a7d6c6-3976-46bc-971b-a67ca5611777 for instance with vm_state building and task_state spawning. [ 819.507608] env[62066]: INFO nova.compute.manager [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] instance snapshotting [ 819.511020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6561640-d591-42b5-be92-37b741b0adb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.533547] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53539213-1b6c-48d7-a813-21a80a9edfca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.668325] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Releasing lock "refresh_cache-4723be94-3479-4e66-8088-914824c0e669" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.669286] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-vif-unplugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.669862] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.670106] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.670573] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.670573] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] No waiting events found dispatching network-vif-unplugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.670773] env[62066]: WARNING nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received unexpected event network-vif-unplugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d for instance with vm_state shelved_offloaded and task_state None. [ 819.670868] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.671067] env[62066]: DEBUG nova.compute.manager [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing instance network info cache due to event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.671273] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.671417] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.671578] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.674828] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 819.675121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.595s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.675470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.651s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.675715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.678024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.310s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.678267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.680871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.696s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.681133] env[62066]: DEBUG nova.objects.instance [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 819.684128] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 819.684128] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 819.701172] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.710392] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Successfully updated port: 03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.730271] env[62066]: INFO nova.scheduler.client.report [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Deleted allocations for instance e95ee795-603e-4cbf-bcd6-1ba54b62a281 [ 819.734866] env[62066]: INFO nova.scheduler.client.report [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleted allocations for instance 2f186d0a-91a0-4dc2-83bd-511099445af7 [ 819.826286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.865609] env[62066]: DEBUG oslo_vmware.api [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.493753} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.866550] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.866801] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.867025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.867268] env[62066]: INFO nova.compute.manager [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Took 1.12 seconds to destroy the instance on the hypervisor. [ 819.867547] env[62066]: DEBUG oslo.service.loopingcall [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.868067] env[62066]: DEBUG nova.compute.manager [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.868185] env[62066]: DEBUG nova.network.neutron [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 820.051595] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 820.052014] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-61d1f5a2-eb8e-4a6f-bed1-17c1db49f069 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.065103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.065380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.067181] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 820.067181] env[62066]: value = "task-1156162" [ 820.067181] env[62066]: _type = "Task" [ 820.067181] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.081248] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156162, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.096540] env[62066]: DEBUG nova.compute.manager [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Received event network-changed-03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.096826] env[62066]: DEBUG nova.compute.manager [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Refreshing instance network info cache due to event network-changed-03a7d6c6-3976-46bc-971b-a67ca5611777. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.096943] env[62066]: DEBUG oslo_concurrency.lockutils [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] Acquiring lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.097557] env[62066]: DEBUG oslo_concurrency.lockutils [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] Acquired lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.097557] env[62066]: DEBUG nova.network.neutron [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Refreshing network info cache for port 03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.159716] env[62066]: DEBUG nova.compute.manager [req-b2fa73d8-0878-437b-9cf5-caa45493b1b2 req-cdafb90c-b2ab-4c6e-9616-00f698cc46d1 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-vif-deleted-b2434f55-6f25-4684-82d8-eb2ac7745888 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.160644] env[62066]: INFO nova.compute.manager [req-b2fa73d8-0878-437b-9cf5-caa45493b1b2 req-cdafb90c-b2ab-4c6e-9616-00f698cc46d1 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Neutron deleted interface b2434f55-6f25-4684-82d8-eb2ac7745888; detaching it from the instance and deleting it from the info cache [ 820.161166] env[62066]: DEBUG nova.network.neutron [req-b2fa73d8-0878-437b-9cf5-caa45493b1b2 req-cdafb90c-b2ab-4c6e-9616-00f698cc46d1 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updating instance_info_cache with network_info: [{"id": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "address": "fa:16:3e:3d:b8:dd", "network": {"id": "d3298ad1-f702-4e3e-b2e4-e2a40d7bbb59", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1342438120", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "00dcb6ace2714cd4994a5273586c5b95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2aee5757-0a", "ovs_interfaceid": "2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.162687] env[62066]: WARNING oslo_messaging._drivers.amqpdriver [req-b2fa73d8-0878-437b-9cf5-caa45493b1b2 req-cdafb90c-b2ab-4c6e-9616-00f698cc46d1 service nova] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 820.199537] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] There are 15 instances to clean {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 820.199537] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: c6fad95f-9167-4ad3-9c1c-16425fa1d36c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 820.209468] env[62066]: DEBUG oslo_vmware.api [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Task: {'id': task-1156159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.641744} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.210371] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.210371] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.210371] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.211097] env[62066]: INFO nova.compute.manager [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] [instance: 4723be94-3479-4e66-8088-914824c0e669] Took 1.67 seconds to destroy the instance on the hypervisor. [ 820.211097] env[62066]: DEBUG oslo.service.loopingcall [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.211097] env[62066]: DEBUG nova.compute.manager [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.211225] env[62066]: DEBUG nova.network.neutron [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 820.213628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.248024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cd3311a2-f286-4c03-8c99-a42d37458bb9 tempest-ServersAaction247Test-457910374 tempest-ServersAaction247Test-457910374-project-member] Lock "e95ee795-603e-4cbf-bcd6-1ba54b62a281" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.971s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.250371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90122f65-6f94-4244-9f8d-391e86ab7351 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "2f186d0a-91a0-4dc2-83bd-511099445af7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.023s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.359280] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.359600] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-031a2ad2-532b-4ac1-8bba-0ea295d1a949 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.367158] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 820.367158] env[62066]: value = "task-1156163" [ 820.367158] env[62066]: _type = "Task" [ 820.367158] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.380048] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156163, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.181449] env[62066]: DEBUG nova.network.neutron [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.182820] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.189254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-24878283-e3a0-43d8-9bbe-04582e478d5b tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.508s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.190996] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 92b150af-351a-43ab-952a-209225733ac8] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 821.193012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.193274] env[62066]: DEBUG nova.network.neutron [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.197885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.197885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.197885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.197885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.197885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.202183] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f87bb454-4abd-4a4f-94dc-1fb252426e19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.204529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.353s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.204745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.206624] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.850s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.206834] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.208549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.116s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.208757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.210807] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.224s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.211035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.212880] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.617s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.214352] env[62066]: INFO nova.compute.claims [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.225144] env[62066]: INFO nova.compute.manager [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Terminating instance [ 821.226087] env[62066]: DEBUG nova.compute.manager [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.226284] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.227963] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7697d68e-5e78-4f80-813f-b66089e816f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.240078] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156163, 'name': PowerOffVM_Task, 'duration_secs': 0.240061} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.240078] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156162, 'name': CreateSnapshot_Task, 'duration_secs': 0.663166} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.242625] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcab7000-7737-4d2a-a51f-5b727903a031 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.254701] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.254701] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 821.255705] env[62066]: INFO nova.scheduler.client.report [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Deleted allocations for instance d9163e64-309b-4381-8819-15757f83ac2e [ 821.261382] env[62066]: INFO nova.scheduler.client.report [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 43d64f76-7f7a-4b95-b9df-c95218612998 [ 821.267427] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5d35fc-90a2-4251-8471-567b06945930 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.271428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47addd6-50cd-4a9a-b7fe-bd28b0494f4f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.271838] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.278043] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71fa071c-27e8-41be-bc0e-db6f62361998 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.280638] env[62066]: INFO nova.scheduler.client.report [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted allocations for instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 [ 821.298213] env[62066]: DEBUG nova.compute.manager [req-b2fa73d8-0878-437b-9cf5-caa45493b1b2 req-cdafb90c-b2ab-4c6e-9616-00f698cc46d1 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Detach interface failed, port_id=b2434f55-6f25-4684-82d8-eb2ac7745888, reason: Instance f196adc6-86bb-4556-bf3d-c52b9108c000 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 821.303710] env[62066]: INFO nova.scheduler.client.report [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Deleted allocations for instance 82e2a71a-d27f-4db9-8f84-16762d3d3bf6 [ 821.325177] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 821.325177] env[62066]: value = "task-1156165" [ 821.325177] env[62066]: _type = "Task" [ 821.325177] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.328889] env[62066]: DEBUG nova.network.neutron [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.333556] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645ec494-8d80-4b52-aa01-3cfa5d128d8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.356443] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.386667] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.387745] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbb131d0-b1df-4698-b419-4ab55c6fd05f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.397200] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 821.397200] env[62066]: value = "task-1156166" [ 821.397200] env[62066]: _type = "Task" [ 821.397200] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.405963] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 821.406226] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.406476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.406628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.406811] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.407075] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-159528b0-dae9-4250-9bbb-50cfcbe4bcc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.418601] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.418815] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.419587] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f71832e-6a86-4a29-a3a3-cbaedf92f6f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.427548] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 821.427548] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521502af-98ca-c7d6-320a-2f9f6cc24d30" [ 821.427548] env[62066]: _type = "Task" [ 821.427548] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.436536] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521502af-98ca-c7d6-320a-2f9f6cc24d30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.483744] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updated VIF entry in instance network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.484161] env[62066]: DEBUG nova.network.neutron [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4991ab48-c8", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.489114] env[62066]: DEBUG nova.network.neutron [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.698939] env[62066]: INFO nova.compute.manager [-] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Took 2.52 seconds to deallocate network for instance. [ 821.699225] env[62066]: DEBUG nova.network.neutron [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.700456] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 3d001331-7a41-4385-a622-54ce481509a8] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 821.704741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.779233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fb410a3-4e81-4a0d-9f5a-fb42890ba89c tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "43d64f76-7f7a-4b95-b9df-c95218612998" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.701s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.782451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db9928e9-9d7a-4903-b2b0-20bc275c2b59 tempest-ServerShowV254Test-2080892666 tempest-ServerShowV254Test-2080892666-project-member] Lock "d9163e64-309b-4381-8819-15757f83ac2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.550s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.790612] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf908b3a-9ada-4785-9b27-4748112214d7 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.683s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.791596] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Acquired lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.792579] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9315ea86-9f65-4259-8049-fb91fe19f406 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.804380] env[62066]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 821.804552] env[62066]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62066) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 821.804919] env[62066]: INFO nova.compute.manager [-] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Took 1.94 seconds to deallocate network for instance. [ 821.807739] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2776e80f-4b05-4f0b-bca1-d2bc13fd24ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.820022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b49917a-555c-40e7-b284-eae7bd29d14a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.841132] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 821.841132] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7940d04d-7bd6-443b-8c26-85408b978c60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.851607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99a47055-a77a-4ef8-a9c4-57a0f48747d5 tempest-AttachInterfacesUnderV243Test-969267617 tempest-AttachInterfacesUnderV243Test-969267617-project-member] Lock "82e2a71a-d27f-4db9-8f84-16762d3d3bf6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.351s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.859929] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 821.859929] env[62066]: value = "task-1156167" [ 821.859929] env[62066]: _type = "Task" [ 821.859929] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.860263] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156165, 'name': PowerOffVM_Task, 'duration_secs': 0.236376} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.872365] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.872615] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 821.875029] env[62066]: ERROR root [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-251607' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-251607' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-251607' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-251607'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-251607' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-251607' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-251607'}\n"]: nova.exception.InstanceNotFound: Instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 could not be found. [ 821.875250] env[62066]: DEBUG oslo_concurrency.lockutils [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] Releasing lock "26d87a85-0aa3-49b9-97ca-1b7fedbebb14" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.875508] env[62066]: DEBUG nova.compute.manager [req-99494278-0af2-479e-a12c-3c29f2245270 req-a709c32c-557b-4c23-b5e7-9291bbe5d349 service nova] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Detach interface failed, port_id=8a5014e0-3b5b-46a1-b1b6-b98e168c945f, reason: Instance 26d87a85-0aa3-49b9-97ca-1b7fedbebb14 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 821.879129] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dec752a1-1b73-4507-af0c-49dc9b1d425d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.886776] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156167, 'name': CloneVM_Task} progress is 12%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.940663] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521502af-98ca-c7d6-320a-2f9f6cc24d30, 'name': SearchDatastore_Task, 'duration_secs': 0.016449} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.943016] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fbef64b-0099-4419-bb42-3f62589f80b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.946221] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.946577] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.946805] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleting the datastore file [datastore2] 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.947155] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8229f666-f863-4483-9f30-83ccfadd47a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.953195] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 821.953195] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52120302-3d11-5155-5b53-286cfaef2297" [ 821.953195] env[62066]: _type = "Task" [ 821.953195] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.954666] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for the task: (returnval){ [ 821.954666] env[62066]: value = "task-1156169" [ 821.954666] env[62066]: _type = "Task" [ 821.954666] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.969567] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52120302-3d11-5155-5b53-286cfaef2297, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.973958] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.988705] env[62066]: DEBUG oslo_concurrency.lockutils [req-3323e0f8-9cf9-4aca-aa3a-783e3ad51dcf req-0747d83c-be86-4845-bd45-04d3d3a2c506 service nova] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.991585] env[62066]: DEBUG oslo_concurrency.lockutils [req-49014cdb-9c64-48a4-8e9d-2de9c92c889a req-88475537-2ea2-494e-83f4-d39ac7859494 service nova] Releasing lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.992043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.992769] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.184593] env[62066]: DEBUG nova.compute.manager [req-02323d52-9dbf-442d-a358-cb192db8fa93 req-5e209926-3bf2-4a15-95bf-a97d774a6544 service nova] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Received event network-vif-deleted-9c31bd1b-5ec5-4e12-bb29-9dc6013f7acf {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.204213] env[62066]: INFO nova.compute.manager [-] [instance: 4723be94-3479-4e66-8088-914824c0e669] Took 1.99 seconds to deallocate network for instance. [ 822.207192] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 1c694323-616c-43b1-bd49-bba707e0788f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 822.215481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.217274] env[62066]: DEBUG nova.compute.manager [req-024b4577-57b1-4ec2-b530-f5ede10304f6 req-6a36c883-1e55-4277-88c4-6616d77d6b83 service nova] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Received event network-vif-deleted-2aee5757-0a9b-4900-a09c-bb1dc5ee5ff6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.217536] env[62066]: DEBUG nova.compute.manager [req-024b4577-57b1-4ec2-b530-f5ede10304f6 req-6a36c883-1e55-4277-88c4-6616d77d6b83 service nova] [instance: 4723be94-3479-4e66-8088-914824c0e669] Received event network-vif-deleted-4d1d88f2-e881-41ed-8de8-7cf13954abf9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.314258] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.383180] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156167, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.469294] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52120302-3d11-5155-5b53-286cfaef2297, 'name': SearchDatastore_Task, 'duration_secs': 0.021061} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.471184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.471540] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 822.471937] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e1a8ef9-97a9-4c89-8751-73b52ae3dac7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.479064] env[62066]: DEBUG oslo_vmware.api [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Task: {'id': task-1156169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231119} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.479064] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.479064] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 822.479064] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.479529] env[62066]: INFO nova.compute.manager [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 1.25 seconds to destroy the instance on the hypervisor. [ 822.479529] env[62066]: DEBUG oslo.service.loopingcall [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.479596] env[62066]: DEBUG nova.compute.manager [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.480066] env[62066]: DEBUG nova.network.neutron [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.484199] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 822.484199] env[62066]: value = "task-1156170" [ 822.484199] env[62066]: _type = "Task" [ 822.484199] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.498095] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156170, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.585294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4216104-0d27-4db0-9699-16b5878825d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.588589] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.595790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff727fe3-892d-46a9-98f6-d07e08794345 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.638425] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9a8371-29d4-409d-8fc9-55b63c732d30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.646845] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ab2dc1-7450-45de-8fdb-1ac9f30b7775 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.661048] env[62066]: DEBUG nova.compute.provider_tree [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.719731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.720176] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 248f5281-eb46-4be1-8642-28813c4b2622] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 822.886747] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156167, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.993831] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156170, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.154236] env[62066]: DEBUG nova.network.neutron [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Updating instance_info_cache with network_info: [{"id": "03a7d6c6-3976-46bc-971b-a67ca5611777", "address": "fa:16:3e:a9:be:66", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03a7d6c6-39", "ovs_interfaceid": "03a7d6c6-3976-46bc-971b-a67ca5611777", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.167069] env[62066]: DEBUG nova.scheduler.client.report [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.223696] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 16c682f7-4aea-4a3a-b22d-ffb5f2b6e0d2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 823.385670] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156167, 'name': CloneVM_Task, 'duration_secs': 1.505676} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.386015] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Created linked-clone VM from snapshot [ 823.386800] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139df6e6-090b-4770-90df-294a157ebf7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.398072] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Uploading image b5438b44-e139-4d11-80c8-d6bdf1282d96 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 823.434060] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 823.434060] env[62066]: value = "vm-251694" [ 823.434060] env[62066]: _type = "VirtualMachine" [ 823.434060] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 823.434873] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-330a911f-ca24-4c82-9750-1d089baa90b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.443591] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease: (returnval){ [ 823.443591] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c34794-cb24-441c-8b8c-0b8a834a0f0c" [ 823.443591] env[62066]: _type = "HttpNfcLease" [ 823.443591] env[62066]: } obtained for exporting VM: (result){ [ 823.443591] env[62066]: value = "vm-251694" [ 823.443591] env[62066]: _type = "VirtualMachine" [ 823.443591] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 823.443849] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the lease: (returnval){ [ 823.443849] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c34794-cb24-441c-8b8c-0b8a834a0f0c" [ 823.443849] env[62066]: _type = "HttpNfcLease" [ 823.443849] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 823.455527] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 823.455527] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c34794-cb24-441c-8b8c-0b8a834a0f0c" [ 823.455527] env[62066]: _type = "HttpNfcLease" [ 823.455527] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 823.499417] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156170, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.501175] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 823.501910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28af584a-d3f7-4b77-9e96-ef208289a481 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.537916] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.538774] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0dcf36a-14e8-4393-9837-65b49365b9b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.570051] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 823.570051] env[62066]: value = "task-1156173" [ 823.570051] env[62066]: _type = "Task" [ 823.570051] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.581091] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156173, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.656829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "refresh_cache-5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.657718] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance network_info: |[{"id": "03a7d6c6-3976-46bc-971b-a67ca5611777", "address": "fa:16:3e:a9:be:66", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03a7d6c6-39", "ovs_interfaceid": "03a7d6c6-3976-46bc-971b-a67ca5611777", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 823.657718] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:be:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e02e98f-44ce-42b7-a3ac-4034fae5d127', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03a7d6c6-3976-46bc-971b-a67ca5611777', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 823.667499] env[62066]: DEBUG oslo.service.loopingcall [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.667669] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 823.667927] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b05cc424-914a-454e-a87a-61d26f252c82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.689875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.690540] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.694733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.805s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.698187] env[62066]: INFO nova.compute.claims [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.709029] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.709029] env[62066]: value = "task-1156174" [ 823.709029] env[62066]: _type = "Task" [ 823.709029] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.722745] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156174, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.728086] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 641bf537-76d2-4c41-a6dc-b5b042327477] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 823.762738] env[62066]: DEBUG nova.network.neutron [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.952923] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 823.952923] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c34794-cb24-441c-8b8c-0b8a834a0f0c" [ 823.952923] env[62066]: _type = "HttpNfcLease" [ 823.952923] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 823.953318] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 823.953318] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c34794-cb24-441c-8b8c-0b8a834a0f0c" [ 823.953318] env[62066]: _type = "HttpNfcLease" [ 823.953318] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 823.955422] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33884a8-620b-4265-97c4-4d30e01cb6de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.965855] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 823.966146] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 824.085868] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156173, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.138670] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-45782a2d-f813-4df9-a732-0c2ab05de45f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.215619] env[62066]: DEBUG nova.compute.utils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 824.221054] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.221291] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.232120] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6a498481-b2ca-4813-87b7-2f09dfa107f4] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 824.242243] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156174, 'name': CreateVM_Task, 'duration_secs': 0.502301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.244029] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.244029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.244029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.245254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.246785] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b6f0216-2f5a-4303-ab28-7e2c127f7b38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.256838] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 824.256838] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528c7277-59e1-6fc7-c033-8edff4f3674f" [ 824.256838] env[62066]: _type = "Task" [ 824.256838] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.267479] env[62066]: INFO nova.compute.manager [-] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Took 1.79 seconds to deallocate network for instance. [ 824.281311] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528c7277-59e1-6fc7-c033-8edff4f3674f, 'name': SearchDatastore_Task, 'duration_secs': 0.014522} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.281968] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.282232] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.282468] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.282613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.282792] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.283396] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-848c4368-d2c4-453a-ba3a-8480aaee5554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.300326] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.300792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.301777] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fdf238a-e266-426f-b972-924907ddd553 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.311335] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 824.311335] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5295e105-440a-2752-5802-fbc9ffbfcda7" [ 824.311335] env[62066]: _type = "Task" [ 824.311335] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.326171] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5295e105-440a-2752-5802-fbc9ffbfcda7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.330342] env[62066]: DEBUG nova.compute.manager [req-1ed72caa-f09b-4882-9520-49af928e15dc req-0679d873-60d6-4d41-94c4-2b5912e87b1e service nova] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Received event network-vif-deleted-ef7a5281-027a-42e2-b91c-21fc55677027 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.340677] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.340826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.375898] env[62066]: DEBUG nova.policy [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91585bdab75c42178c835de912fd3fef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08737e043ede43b58298d1c1a834fa84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.585166] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156173, 'name': ReconfigVM_Task, 'duration_secs': 0.609622} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.585467] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfigured VM instance instance-00000045 to attach disk [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.586442] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53b370c-93c7-46fd-a043-e284df43a12b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.624616] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5351dd4-292d-4f4b-8af8-86d280d2ae11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.646911] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 824.646911] env[62066]: value = "task-1156175" [ 824.646911] env[62066]: _type = "Task" [ 824.646911] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.655749] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.726708] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.743870] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 265906be-704e-4f0f-ae33-e89135e27268] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 824.783525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.829674] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5295e105-440a-2752-5802-fbc9ffbfcda7, 'name': SearchDatastore_Task, 'duration_secs': 0.015381} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.831154] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-680a103c-6c92-451e-857a-84219673e391 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.841097] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 824.841097] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d3bdcf-e980-0f7d-f840-655baba4242a" [ 824.841097] env[62066]: _type = "Task" [ 824.841097] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.848322] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.867683] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d3bdcf-e980-0f7d-f840-655baba4242a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.155890] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156175, 'name': ReconfigVM_Task, 'duration_secs': 0.272786} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.156254] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.157118] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8874995-d13d-47d0-986c-569aec440658 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.160411] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2909e9-9b52-4850-91e6-d1b4a63449c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.167727] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 825.167727] env[62066]: value = "task-1156176" [ 825.167727] env[62066]: _type = "Task" [ 825.167727] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.174608] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ea2326-8706-4e51-9984-a0d61a98c5f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.179334] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Successfully created port: 94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.214721] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee1999b-e3d1-489d-bc17-79f49a617b7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.218052] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156176, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.225354] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f196a35-605e-4812-8791-5fb6ca145485 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.242610] env[62066]: DEBUG nova.compute.provider_tree [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.251223] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: c0ac4362-766f-48ba-aeb2-7fd976c1f47f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 825.355773] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d3bdcf-e980-0f7d-f840-655baba4242a, 'name': SearchDatastore_Task, 'duration_secs': 0.020041} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.356175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.356555] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4/5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.356914] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0af0e0c3-78ce-463d-80f7-3947a8f598b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.369948] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 825.369948] env[62066]: value = "task-1156177" [ 825.369948] env[62066]: _type = "Task" [ 825.369948] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.382727] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.386364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.684412] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156176, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.746234] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.751159] env[62066]: DEBUG nova.scheduler.client.report [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.755079] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d6837004-9b24-45f7-9ed7-48c4503859cb] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 825.888629] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156177, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.915720] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.915720] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.915935] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.916082] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.916247] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.916901] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.916901] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.916901] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.917097] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.917279] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.917445] env[62066]: DEBUG nova.virt.hardware [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.922026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074a7f43-285b-416c-bb5f-246f66bc0f99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.937224] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 825.939640] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee568153-fb34-416b-80a5-8a807469e736 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.947278] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dddd76-ecee-46e9-a7f8-8c8561c666d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.968422] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 825.968854] env[62066]: ERROR oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk due to incomplete transfer. [ 825.969300] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-60163056-5114-4a4e-941d-26ad4fac0c5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.981840] env[62066]: DEBUG oslo_vmware.rw_handles [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520155ac-a658-16a1-c322-886290b9cf31/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 825.982114] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Uploaded image 7edbe552-3036-4d5b-a9e5-62e23f253c7a to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 825.984760] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 825.985201] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4c5be6b5-41d2-47cc-8b6a-1aa81c212651 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.993574] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 825.993574] env[62066]: value = "task-1156178" [ 825.993574] env[62066]: _type = "Task" [ 825.993574] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.006326] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156178, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.188125] env[62066]: DEBUG oslo_vmware.api [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156176, 'name': PowerOnVM_Task, 'duration_secs': 0.536196} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.188125] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.189378] env[62066]: DEBUG nova.compute.manager [None req-867e6056-2126-42a4-9a7e-b17dd6df162a tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.190319] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d94cf6-a042-418a-8977-f399e638fb42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.262020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.262020] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.264298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.541s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.264693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.267095] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.407s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.267443] env[62066]: DEBUG nova.objects.instance [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 826.270208] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: da7e4be3-29fc-462c-b8c7-ce4d80d19a28] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 826.383265] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651024} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.383518] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4/5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.383767] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.384079] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aee0b166-24e4-49f5-b619-5979c0701244 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.393267] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 826.393267] env[62066]: value = "task-1156179" [ 826.393267] env[62066]: _type = "Task" [ 826.393267] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.404057] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156179, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.512350] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156178, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.774494] env[62066]: DEBUG nova.compute.utils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 826.789409] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 2dc8d357-2864-410a-93ef-21ea4e9b530c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 826.799117] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.799815] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.804383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d19fd86e-21e2-416e-be5c-20a3905cb941 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.215s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.805853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.613s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.806134] env[62066]: INFO nova.compute.manager [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Unshelving [ 826.886393] env[62066]: DEBUG nova.policy [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6b41a7e94574415be138af85d4b819d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e25052e3fae44617b5a63f881a7b081b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.908019] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156179, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.908019] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.908019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a000b5bd-919a-416f-a92c-2cf323af2cc2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.933758] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4/5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.934151] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b8a1f72-ef6c-44ea-a643-709144de5a43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.958114] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 826.958114] env[62066]: value = "task-1156180" [ 826.958114] env[62066]: _type = "Task" [ 826.958114] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.979066] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156180, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.010827] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156178, 'name': Destroy_Task, 'duration_secs': 0.716033} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.011275] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroyed the VM [ 827.012192] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 827.012192] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-11ef5151-6b27-46a5-b438-d08797231d51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.021027] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 827.021027] env[62066]: value = "task-1156181" [ 827.021027] env[62066]: _type = "Task" [ 827.021027] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.032178] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156181, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.297921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb2c4fdf-5f3f-4ad0-90cf-2e56c5a13429 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.031s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.299476] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.305039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.600s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.308100] env[62066]: INFO nova.compute.claims [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.312468] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: b4cb3d6d-d8ec-4632-8020-fae39b18143e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 827.475494] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156180, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.537078] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156181, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.634534] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Successfully created port: f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.731494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.731830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.738552] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Successfully updated port: 94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.821195] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: fa73770f-784a-44c3-ba65-e7176180760b] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 827.846816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.975448] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156180, 'name': ReconfigVM_Task, 'duration_secs': 0.593912} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.976372] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4/5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.977229] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48688028-6688-41af-8fed-b5e5025e8892 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.989520] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 827.989520] env[62066]: value = "task-1156182" [ 827.989520] env[62066]: _type = "Task" [ 827.989520] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.004891] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156182, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.036828] env[62066]: DEBUG oslo_vmware.api [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156181, 'name': RemoveSnapshot_Task, 'duration_secs': 0.839462} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.037251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 828.037971] env[62066]: INFO nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 15.64 seconds to snapshot the instance on the hypervisor. [ 828.244693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.244693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.244693] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.244693] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.328355] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.331699] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 828.331699] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances with incomplete migration {{(pid=62066) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 828.356316] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.356608] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.356821] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.356972] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.358877] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.359247] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.360861] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.361109] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.361322] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.361534] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.361793] env[62066]: DEBUG nova.virt.hardware [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.365219] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec36d31-f1c0-4a09-8f6f-510dc16c4a9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.379349] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebbeecb-ebc4-464d-b58c-dd049e4fb069 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.502020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.502461] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.503682] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156182, 'name': Rename_Task, 'duration_secs': 0.257512} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.509292] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.509292] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cee1bb1-caaf-46d4-a76c-71f79c589ca6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.518582] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 828.518582] env[62066]: value = "task-1156183" [ 828.518582] env[62066]: _type = "Task" [ 828.518582] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.536229] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.597114] env[62066]: DEBUG nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Found 3 images (rotation: 2) {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 828.597507] env[62066]: DEBUG nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Rotating out 1 backups {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 828.597779] env[62066]: DEBUG nova.compute.manager [None req-d58bf054-375d-4da8-aa0e-0e86eca4278d tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleting image 63c467a0-7d85-4619-a71e-fe2e06314300 {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 828.766411] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef0a6d0-95f1-49d3-a2f5-1d426aabd929 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.776937] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.778169] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb492f3-c9fd-4dcb-a92b-b8d9d518ad6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.824754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7547f1b9-d8ba-40bb-b5ce-29a3ef2e85fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.835711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad39d5ae-758d-411e-89c4-81bec6dfd73d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.841727] env[62066]: DEBUG nova.compute.manager [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Received event network-vif-plugged-94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.841727] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.841727] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.841727] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.841727] env[62066]: DEBUG nova.compute.manager [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] No waiting events found dispatching network-vif-plugged-94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.842274] env[62066]: WARNING nova.compute.manager [req-0c1c3f81-ca8d-4708-b13a-c03ec7d2897c req-42de289b-e53c-4b33-bf55-28c85b31603e service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Received unexpected event network-vif-plugged-94c39e5f-27a8-4373-9893-8defce990708 for instance with vm_state building and task_state spawning. [ 828.846153] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.848196] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 828.868377] env[62066]: DEBUG nova.compute.provider_tree [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.004657] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.031413] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.099666] env[62066]: DEBUG nova.network.neutron [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Updating instance_info_cache with network_info: [{"id": "94c39e5f-27a8-4373-9893-8defce990708", "address": "fa:16:3e:a5:14:cf", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c39e5f-27", "ovs_interfaceid": "94c39e5f-27a8-4373-9893-8defce990708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.374803] env[62066]: DEBUG nova.scheduler.client.report [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.383747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.384021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.385073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.385299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.491874] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Successfully updated port: f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.533324] env[62066]: DEBUG oslo_vmware.api [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156183, 'name': PowerOnVM_Task, 'duration_secs': 0.718183} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.534120] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.534120] env[62066]: INFO nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Took 11.40 seconds to spawn the instance on the hypervisor. [ 829.534120] env[62066]: DEBUG nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.535572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c66385f-4098-4ddf-8029-8b32788c5a00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.538899] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.587018] env[62066]: INFO nova.compute.manager [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Unrescuing [ 829.587365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.587520] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquired lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.587688] env[62066]: DEBUG nova.network.neutron [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.602930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.603302] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance network_info: |[{"id": "94c39e5f-27a8-4373-9893-8defce990708", "address": "fa:16:3e:a5:14:cf", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c39e5f-27", "ovs_interfaceid": "94c39e5f-27a8-4373-9893-8defce990708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.603990] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:14:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94c39e5f-27a8-4373-9893-8defce990708', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.611563] env[62066]: DEBUG oslo.service.loopingcall [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.611829] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.612452] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e653b71b-d867-4b7d-9c9d-388a9c86bf2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.638120] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.638120] env[62066]: value = "task-1156184" [ 829.638120] env[62066]: _type = "Task" [ 829.638120] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.647897] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156184, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.885085] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.885869] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.888851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.673s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.889266] env[62066]: DEBUG nova.objects.instance [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lazy-loading 'resources' on Instance uuid f196adc6-86bb-4556-bf3d-c52b9108c000 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.893980] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.894296] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.994617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.994866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquired lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.995110] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.056838] env[62066]: INFO nova.compute.manager [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Took 37.90 seconds to build instance. [ 830.154385] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156184, 'name': CreateVM_Task, 'duration_secs': 0.407281} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.154577] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.155298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.155471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.155826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.156183] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee074fb-73c2-46d3-92ae-bebc64f52050 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.164252] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 830.164252] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52503c54-859d-b141-0881-c761a4bcb30a" [ 830.164252] env[62066]: _type = "Task" [ 830.164252] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.176560] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52503c54-859d-b141-0881-c761a4bcb30a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.348058] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 830.348327] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 830.348527] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 830.403136] env[62066]: DEBUG nova.compute.utils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 830.412657] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.412845] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 830.429076] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.439930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.472770] env[62066]: DEBUG nova.network.neutron [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updating instance_info_cache with network_info: [{"id": "6383bfb0-3350-4704-bc27-2450027edc30", "address": "fa:16:3e:c4:b2:40", "network": {"id": "4c93a890-6946-4c34-b09d-e5762bcd2fcd", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-862238786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8a8e27dbc37043b3b36f8492c47d2286", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6383bfb0-33", "ovs_interfaceid": "6383bfb0-3350-4704-bc27-2450027edc30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.497280] env[62066]: DEBUG nova.policy [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85a0637645c443a9d9e6028053c0450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '350ba3c5676a4dd0a018900e7237a5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.557927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfdb7348-d64e-40e8-9277-742e665dc0a4 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.411s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.565748] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.677280] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52503c54-859d-b141-0881-c761a4bcb30a, 'name': SearchDatastore_Task, 'duration_secs': 0.020882} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.677428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.677709] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.677983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.678173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.678365] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.681608] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a516719b-7f7e-452c-8a74-76e76220674e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.692172] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.693929] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.693929] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3fab666-6fd4-4afa-87bc-0464b489a9cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.700656] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 830.700656] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5228ddc8-7aa0-c546-4c56-e89c57447bf3" [ 830.700656] env[62066]: _type = "Task" [ 830.700656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.713394] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5228ddc8-7aa0-c546-4c56-e89c57447bf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.850764] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2258ef9-76b1-4034-925e-6eafa354ce60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.860296] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81475eb-4553-480a-916c-4139bacfb9ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.899093] env[62066]: DEBUG nova.network.neutron [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updating instance_info_cache with network_info: [{"id": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "address": "fa:16:3e:59:33:1f", "network": {"id": "fd494499-759b-4d73-a332-60a6f9a126af", "bridge": "br-int", "label": "tempest-ServersTestJSON-207731434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25052e3fae44617b5a63f881a7b081b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19a58d1-80", "ovs_interfaceid": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.903997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffd3c61-45a4-4555-858f-2635bf85b843 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.911678] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.920555] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e453a376-1a8f-4ddd-9be2-2b3870f40d1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.936814] env[62066]: DEBUG nova.compute.provider_tree [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.975889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Releasing lock "refresh_cache-f3ef43e8-6092-44ac-8990-979810a9748f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.976697] env[62066]: DEBUG nova.objects.instance [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lazy-loading 'flavor' on Instance uuid f3ef43e8-6092-44ac-8990-979810a9748f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.996540] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Successfully created port: 3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.095050] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Received event network-changed-94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.095288] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Refreshing instance network info cache due to event network-changed-94c39e5f-27a8-4373-9893-8defce990708. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.095515] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Acquiring lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.095659] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Acquired lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.095868] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Refreshing network info cache for port 94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.213074] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5228ddc8-7aa0-c546-4c56-e89c57447bf3, 'name': SearchDatastore_Task, 'duration_secs': 0.018482} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.213917] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-750c0fd7-da62-417f-adad-814cd63c9acf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.221324] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 831.221324] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5290ed6f-73ee-37b8-181d-6c43c4fa8d9e" [ 831.221324] env[62066]: _type = "Task" [ 831.221324] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.231989] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5290ed6f-73ee-37b8-181d-6c43c4fa8d9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.386094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.386094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.386094] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 831.404943] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Releasing lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.405291] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Instance network_info: |[{"id": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "address": "fa:16:3e:59:33:1f", "network": {"id": "fd494499-759b-4d73-a332-60a6f9a126af", "bridge": "br-int", "label": "tempest-ServersTestJSON-207731434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25052e3fae44617b5a63f881a7b081b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19a58d1-80", "ovs_interfaceid": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.405961] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:33:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f19a58d1-80e6-4f51-bd80-36927b0c3867', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.415254] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Creating folder: Project (e25052e3fae44617b5a63f881a7b081b). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.415854] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c2f0f0a-8a1b-4196-9d11-7ab436abd773 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.429463] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Created folder: Project (e25052e3fae44617b5a63f881a7b081b) in parent group-v251573. [ 831.429665] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Creating folder: Instances. Parent ref: group-v251697. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.429911] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f574c24-a02c-4419-a155-d0d584a9a2c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.439490] env[62066]: DEBUG nova.scheduler.client.report [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.444741] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Created folder: Instances in parent group-v251697. [ 831.444990] env[62066]: DEBUG oslo.service.loopingcall [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.445208] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.445432] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42ffd0b2-e7af-4f80-95d2-1ded217cb055 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.471039] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.471039] env[62066]: value = "task-1156187" [ 831.471039] env[62066]: _type = "Task" [ 831.471039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.483785] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156187, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.484898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b60be13-3c7f-4dc0-a97e-8ac268c04458 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.513525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.514393] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbb8d0b6-eced-4a30-beeb-933f6b792ef7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.522724] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 831.522724] env[62066]: value = "task-1156188" [ 831.522724] env[62066]: _type = "Task" [ 831.522724] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.534574] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.676972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.677381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.677641] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.677864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.678095] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.680566] env[62066]: INFO nova.compute.manager [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Terminating instance [ 831.685151] env[62066]: DEBUG nova.compute.manager [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.685400] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.686665] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662d12ca-22fc-46cc-9f79-b684eca39b3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.696742] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.697034] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2aa56433-8d27-43fa-b9dc-57332ecba16e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.705442] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 831.705442] env[62066]: value = "task-1156189" [ 831.705442] env[62066]: _type = "Task" [ 831.705442] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.715979] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.732175] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5290ed6f-73ee-37b8-181d-6c43c4fa8d9e, 'name': SearchDatastore_Task, 'duration_secs': 0.025321} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.735434] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.735434] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.735434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2df86cd2-bb41-4c1d-a65d-c9e9acadf160 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.744976] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 831.744976] env[62066]: value = "task-1156190" [ 831.744976] env[62066]: _type = "Task" [ 831.744976] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.754082] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.878597] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Updated VIF entry in instance network info cache for port 94c39e5f-27a8-4373-9893-8defce990708. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.879157] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Updating instance_info_cache with network_info: [{"id": "94c39e5f-27a8-4373-9893-8defce990708", "address": "fa:16:3e:a5:14:cf", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94c39e5f-27", "ovs_interfaceid": "94c39e5f-27a8-4373-9893-8defce990708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.927451] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.956471] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.956791] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.956892] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.957097] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.957256] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.957507] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.958208] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.958497] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.958876] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.959061] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.959338] env[62066]: DEBUG nova.virt.hardware [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.960652] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2893b4-7486-4886-9251-c68a0e3d21aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.966369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.077s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.970109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.656s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.970314] env[62066]: DEBUG nova.objects.instance [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'resources' on Instance uuid b46b7795-b7f7-4c55-bc72-184bb7bb8842 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.982494] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579a8ea6-2dc2-4905-80f0-478157d7dcf0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.993302] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156187, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.007249] env[62066]: INFO nova.scheduler.client.report [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Deleted allocations for instance f196adc6-86bb-4556-bf3d-c52b9108c000 [ 832.033574] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156188, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.219225] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156189, 'name': PowerOffVM_Task, 'duration_secs': 0.236718} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.219488] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.219639] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.220278] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-520cf38b-25c2-4f64-b984-5fa4a8ec81ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.258059] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156190, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.303987] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.304396] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.304542] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleting the datastore file [datastore2] 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.304883] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79ef8d02-159e-4b51-98b8-bc4aafe2da66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.315022] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 832.315022] env[62066]: value = "task-1156192" [ 832.315022] env[62066]: _type = "Task" [ 832.315022] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.324692] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.383374] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Releasing lock "refresh_cache-68bf0e89-8236-44ef-b514-d25698eb23c3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.383374] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Received event network-vif-plugged-f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.383374] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Acquiring lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.383374] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.383374] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.383374] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] No waiting events found dispatching network-vif-plugged-f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.383374] env[62066]: WARNING nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Received unexpected event network-vif-plugged-f19a58d1-80e6-4f51-bd80-36927b0c3867 for instance with vm_state building and task_state spawning. [ 832.383374] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Received event network-changed-f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.383717] env[62066]: DEBUG nova.compute.manager [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Refreshing instance network info cache due to event network-changed-f19a58d1-80e6-4f51-bd80-36927b0c3867. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.383717] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Acquiring lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.383774] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Acquired lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.383945] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Refreshing network info cache for port f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.493022] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156187, 'name': CreateVM_Task, 'duration_secs': 0.559793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.493022] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 832.493022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.493022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.493022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.493022] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3855111-a879-416b-ab87-c91b4e81ef21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.498737] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 832.498737] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522ef18d-6c8c-ac75-5ea2-3b9be48f2fb6" [ 832.498737] env[62066]: _type = "Task" [ 832.498737] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.507096] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522ef18d-6c8c-ac75-5ea2-3b9be48f2fb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.522704] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e820503-f2c6-4b5b-a567-d55d98659626 tempest-ServersTestMultiNic-909033266 tempest-ServersTestMultiNic-909033266-project-member] Lock "f196adc6-86bb-4556-bf3d-c52b9108c000" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.604s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.539217] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156188, 'name': PowerOffVM_Task, 'duration_secs': 0.52676} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.539664] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.549401] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 832.555500] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85994323-f86a-44b1-a6fc-0fe54b27949d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.581123] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 832.581123] env[62066]: value = "task-1156193" [ 832.581123] env[62066]: _type = "Task" [ 832.581123] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.593191] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156193, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.729475] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Successfully updated port: 3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.766649] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681968} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.771025] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.771025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.771025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c10b79b0-3525-4947-b46b-7e76c1b2984b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.784356] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 832.784356] env[62066]: value = "task-1156194" [ 832.784356] env[62066]: _type = "Task" [ 832.784356] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.797400] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updating instance_info_cache with network_info: [{"id": "59b6e38f-b186-420c-89c8-c860e8310108", "address": "fa:16:3e:5c:97:e8", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b6e38f-b1", "ovs_interfaceid": "59b6e38f-b186-420c-89c8-c860e8310108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.800355] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.825738] env[62066]: DEBUG oslo_vmware.api [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.38079} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.828667] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.828895] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.829089] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.829284] env[62066]: INFO nova.compute.manager [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 832.829566] env[62066]: DEBUG oslo.service.loopingcall [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.830455] env[62066]: DEBUG nova.compute.manager [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.830565] env[62066]: DEBUG nova.network.neutron [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.992029] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8128c75-9fcf-4022-b907-8dcef89b18f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.007759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4f23fc-8f74-4035-b9df-c524ff7c9202 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.015142] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522ef18d-6c8c-ac75-5ea2-3b9be48f2fb6, 'name': SearchDatastore_Task, 'duration_secs': 0.023898} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.015963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.016399] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.017133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.017133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.017133] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.017133] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a3a37f0-c5f5-4480-9f33-1e0c3481bd05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.054456] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04633b3a-061d-4349-a951-a875725551e4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.062163] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.062216] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.066419] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c29861d-c2f4-48b7-9a91-edcd9d549fc2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.070097] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2561fe9-90ab-40d3-a626-2f2aaa52215a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.087041] env[62066]: DEBUG nova.compute.provider_tree [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.095733] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 833.095733] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529457e6-0050-5b9b-c33a-d0ef78293ad7" [ 833.095733] env[62066]: _type = "Task" [ 833.095733] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.105649] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156193, 'name': ReconfigVM_Task, 'duration_secs': 0.242832} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.108032] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 833.108032] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.108354] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a05eb4f-6e64-458d-9112-1e011e4e6bde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.115023] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529457e6-0050-5b9b-c33a-d0ef78293ad7, 'name': SearchDatastore_Task, 'duration_secs': 0.016696} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.115375] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bf9ae97-dff1-477b-ade7-5ef7950a331b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.119942] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 833.119942] env[62066]: value = "task-1156195" [ 833.119942] env[62066]: _type = "Task" [ 833.119942] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.121337] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 833.121337] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5268557b-f786-543d-bf56-097384de18ce" [ 833.121337] env[62066]: _type = "Task" [ 833.121337] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.134134] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.140173] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268557b-f786-543d-bf56-097384de18ce, 'name': SearchDatastore_Task, 'duration_secs': 0.01073} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.140819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.141213] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3092f591-214f-40ba-a8e2-ccc6a0f007c9/3092f591-214f-40ba-a8e2-ccc6a0f007c9.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.142303] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51705dc9-8082-458b-8e9a-c92ff01c8870 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.153814] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 833.153814] env[62066]: value = "task-1156196" [ 833.153814] env[62066]: _type = "Task" [ 833.153814] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.164330] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.210286] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 833.211294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1002432f-861d-4cc2-aa28-160218b2f68d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.219145] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 833.219385] env[62066]: ERROR oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk due to incomplete transfer. [ 833.219647] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e30a6175-6d77-4928-9fbc-da5c5f130ffe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.227759] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d737c9-3d43-5ce6-5725-cf98fa5e6ee2/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 833.227970] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Uploaded image b5438b44-e139-4d11-80c8-d6bdf1282d96 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 833.230287] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 833.230981] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-04275399-8b43-4bf3-867f-dbdfbe5dd1c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.233837] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.233837] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.233837] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.238528] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 833.238528] env[62066]: value = "task-1156197" [ 833.238528] env[62066]: _type = "Task" [ 833.238528] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.252654] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156197, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.297120] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071342} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.297120] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.297120] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3632d9-20db-4931-b377-c24b1ee5c0ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.310702] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.310814] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 833.319516] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.321313] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.322058] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba4dd14b-5260-4429-a97a-e84fa7fe4811 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.337562] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updated VIF entry in instance network info cache for port f19a58d1-80e6-4f51-bd80-36927b0c3867. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.337926] env[62066]: DEBUG nova.network.neutron [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updating instance_info_cache with network_info: [{"id": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "address": "fa:16:3e:59:33:1f", "network": {"id": "fd494499-759b-4d73-a332-60a6f9a126af", "bridge": "br-int", "label": "tempest-ServersTestJSON-207731434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25052e3fae44617b5a63f881a7b081b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19a58d1-80", "ovs_interfaceid": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.339695] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.340143] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.340757] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.341237] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.341460] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.341576] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 833.341722] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.358446] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 833.358446] env[62066]: value = "task-1156198" [ 833.358446] env[62066]: _type = "Task" [ 833.358446] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.368359] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156198, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.410260] env[62066]: DEBUG nova.compute.manager [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Received event network-vif-plugged-3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.410523] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.410744] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.411037] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.411790] env[62066]: DEBUG nova.compute.manager [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] No waiting events found dispatching network-vif-plugged-3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.411970] env[62066]: WARNING nova.compute.manager [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Received unexpected event network-vif-plugged-3ea61126-db59-442a-91ed-a46b37e22462 for instance with vm_state building and task_state spawning. [ 833.412177] env[62066]: DEBUG nova.compute.manager [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Received event network-changed-3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.412338] env[62066]: DEBUG nova.compute.manager [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Refreshing instance network info cache due to event network-changed-3ea61126-db59-442a-91ed-a46b37e22462. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.412529] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Acquiring lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.563474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.563474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.563474] env[62066]: DEBUG nova.compute.manager [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.563474] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156f91a5-a693-40e4-a9a1-cc87a07c3f66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.573509] env[62066]: DEBUG nova.compute.manager [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 833.574517] env[62066]: DEBUG nova.objects.instance [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.597634] env[62066]: DEBUG nova.scheduler.client.report [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.638197] env[62066]: DEBUG oslo_vmware.api [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156195, 'name': PowerOnVM_Task, 'duration_secs': 0.496539} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.638197] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.638197] env[62066]: DEBUG nova.compute.manager [None req-cc67b820-7031-4f93-9350-a5db2363195f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.638629] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2c26f0-a4ef-4620-80e9-9a7ac66e944d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.665473] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156196, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.752086] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156197, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.775365] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.839767] env[62066]: DEBUG nova.network.neutron [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.841452] env[62066]: DEBUG oslo_concurrency.lockutils [req-4dcc4826-1fbd-4992-92ae-ed3bb35bc46f req-353dade7-92eb-4919-bfed-ca3a958a877f service nova] Releasing lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.845856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.870610] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156198, 'name': ReconfigVM_Task, 'duration_secs': 0.407815} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.871305] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.871692] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c579b6d8-c892-4499-9db8-8656cadfbd6d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.880609] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 833.880609] env[62066]: value = "task-1156199" [ 833.880609] env[62066]: _type = "Task" [ 833.880609] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.893437] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156199, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.926311] env[62066]: DEBUG nova.network.neutron [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.082095] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.082568] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60038873-91b8-4d1d-9f61-7118a05cfdbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.093019] env[62066]: DEBUG oslo_vmware.api [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 834.093019] env[62066]: value = "task-1156200" [ 834.093019] env[62066]: _type = "Task" [ 834.093019] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.104803] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.135s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.109123] env[62066]: DEBUG oslo_vmware.api [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.109123] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.388s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.109123] env[62066]: DEBUG nova.objects.instance [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lazy-loading 'resources' on Instance uuid 4723be94-3479-4e66-8088-914824c0e669 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.133794] env[62066]: INFO nova.scheduler.client.report [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocations for instance b46b7795-b7f7-4c55-bc72-184bb7bb8842 [ 834.171040] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558805} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.171040] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3092f591-214f-40ba-a8e2-ccc6a0f007c9/3092f591-214f-40ba-a8e2-ccc6a0f007c9.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.171040] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.171040] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79f36624-c839-4049-a4f5-5a2c2121f15e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.181084] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 834.181084] env[62066]: value = "task-1156201" [ 834.181084] env[62066]: _type = "Task" [ 834.181084] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.191808] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156201, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.254765] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156197, 'name': Destroy_Task, 'duration_secs': 0.640551} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.255090] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Destroyed the VM [ 834.255576] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 834.256296] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-dce54c75-a4c8-448c-986d-2bb49e26c7a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.264415] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 834.264415] env[62066]: value = "task-1156202" [ 834.264415] env[62066]: _type = "Task" [ 834.264415] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.275667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.275951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.286257] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156202, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.344871] env[62066]: INFO nova.compute.manager [-] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Took 1.51 seconds to deallocate network for instance. [ 834.391670] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156199, 'name': Rename_Task, 'duration_secs': 0.154243} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.392678] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 834.392973] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9253e9c0-415a-4327-879d-e6df7d78a52d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.402771] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 834.402771] env[62066]: value = "task-1156203" [ 834.402771] env[62066]: _type = "Task" [ 834.402771] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.413269] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.429185] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.429637] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Instance network_info: |[{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.430460] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Acquired lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.430645] env[62066]: DEBUG nova.network.neutron [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Refreshing network info cache for port 3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.433604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:9a:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ea61126-db59-442a-91ed-a46b37e22462', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.441965] env[62066]: DEBUG oslo.service.loopingcall [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.441965] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.441965] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-343b173e-42d8-4499-9cf0-b07226483a01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.465808] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.465808] env[62066]: value = "task-1156204" [ 834.465808] env[62066]: _type = "Task" [ 834.465808] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.477132] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156204, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.608080] env[62066]: DEBUG oslo_vmware.api [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156200, 'name': PowerOffVM_Task, 'duration_secs': 0.284647} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.608433] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.608671] env[62066]: DEBUG nova.compute.manager [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.610578] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852f11d0-9dff-408e-bbd1-0b7cf2b8a3bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.647998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e661ba9a-373e-4630-b04c-6bd48fe1b0cc tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "b46b7795-b7f7-4c55-bc72-184bb7bb8842" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.907s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.699463] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156201, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073142} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.699463] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.700320] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e66b76a-4fcf-4211-aecd-a49818900db2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.737366] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 3092f591-214f-40ba-a8e2-ccc6a0f007c9/3092f591-214f-40ba-a8e2-ccc6a0f007c9.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.738534] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4a45859-c525-4238-acb5-77fe52026194 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.764547] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 834.764547] env[62066]: value = "task-1156205" [ 834.764547] env[62066]: _type = "Task" [ 834.764547] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.779984] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.790829] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156202, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.791914] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.852134] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.914292] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156203, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.981357] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156204, 'name': CreateVM_Task, 'duration_secs': 0.42966} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.981630] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.982214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.985309] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.985309] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.985309] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c0576f-3c9f-4c91-a413-adb887f6c5ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.990058] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 834.990058] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2ef0c-4894-a26d-8bcd-d2e7bf5ca3de" [ 834.990058] env[62066]: _type = "Task" [ 834.990058] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.000277] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2ef0c-4894-a26d-8bcd-d2e7bf5ca3de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.070045] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7ae66f-4cb1-45b0-9cb9-c1bcd4c2eb0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.078293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f74cea-1107-47f9-b9a0-b068911092b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.109950] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf5fc8c-f329-4f46-b229-761ccf1f2d82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.119392] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfbba76-3047-4984-8b06-87cf619fd4d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.147632] env[62066]: DEBUG nova.compute.provider_tree [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.149697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f57e9e68-c27a-4583-ac8b-8c6843879373 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.589s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.192839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "f3ef43e8-6092-44ac-8990-979810a9748f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.193187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.193517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.193859] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.194193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.197716] env[62066]: INFO nova.compute.manager [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Terminating instance [ 835.204250] env[62066]: DEBUG nova.compute.manager [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.204479] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.205398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bae0cf9-9639-4b25-9532-a908a58bea9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.214760] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.215028] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f09cb356-edd1-4086-8249-36e160c0783b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.221943] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 835.221943] env[62066]: value = "task-1156206" [ 835.221943] env[62066]: _type = "Task" [ 835.221943] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.230739] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156206, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.278393] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.285050] env[62066]: DEBUG oslo_vmware.api [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156202, 'name': RemoveSnapshot_Task, 'duration_secs': 0.7278} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.286023] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 835.286023] env[62066]: INFO nova.compute.manager [None req-4f682739-83ff-492e-9c95-a63e7c3638bd tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Took 15.78 seconds to snapshot the instance on the hypervisor. [ 835.313892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.419100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.419506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.428493] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156203, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.430153] env[62066]: DEBUG nova.network.neutron [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updated VIF entry in instance network info cache for port 3ea61126-db59-442a-91ed-a46b37e22462. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 835.430677] env[62066]: DEBUG nova.network.neutron [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.505367] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b2ef0c-4894-a26d-8bcd-d2e7bf5ca3de, 'name': SearchDatastore_Task, 'duration_secs': 0.009842} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.505367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.505499] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.505745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.506014] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.506128] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.506859] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75af6810-c681-4d95-9305-ba436dd875d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.518243] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.518488] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.519399] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c762fe-bcb1-4344-8d74-00f7688ed699 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.525898] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 835.525898] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d2b24b-1295-6f39-51db-9862b9bcda1b" [ 835.525898] env[62066]: _type = "Task" [ 835.525898] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.535475] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d2b24b-1295-6f39-51db-9862b9bcda1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.543240] env[62066]: DEBUG nova.compute.manager [req-d7d95871-bc28-4b00-a38d-f3cf316c00a1 req-5ca37729-0421-4cd6-afee-4f69b07362b8 service nova] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Received event network-vif-deleted-03a7d6c6-3976-46bc-971b-a67ca5611777 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.655568] env[62066]: DEBUG nova.scheduler.client.report [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.737353] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156206, 'name': PowerOffVM_Task, 'duration_secs': 0.241156} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.737353] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.737353] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.737353] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02adec86-4bab-40f2-bf7e-91f3dfbec784 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.775884] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156205, 'name': ReconfigVM_Task, 'duration_secs': 0.645666} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.776218] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 3092f591-214f-40ba-a8e2-ccc6a0f007c9/3092f591-214f-40ba-a8e2-ccc6a0f007c9.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.776847] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-836992b7-4827-40d9-a0db-7b12a4b24354 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.784623] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 835.784623] env[62066]: value = "task-1156208" [ 835.784623] env[62066]: _type = "Task" [ 835.784623] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.796230] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156208, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.801442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.801663] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.801845] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleting the datastore file [datastore1] f3ef43e8-6092-44ac-8990-979810a9748f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.802163] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b4ec27c-1afa-42d4-98c7-f20dc72fbc52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.811294] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 835.811294] env[62066]: value = "task-1156209" [ 835.811294] env[62066]: _type = "Task" [ 835.811294] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.821567] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156209, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.917382] env[62066]: DEBUG oslo_vmware.api [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156203, 'name': PowerOnVM_Task, 'duration_secs': 1.336034} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.917678] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 835.918407] env[62066]: INFO nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Took 10.17 seconds to spawn the instance on the hypervisor. [ 835.918407] env[62066]: DEBUG nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.919130] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5a8db9-291c-41cf-84e9-ce65eb6aac02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.930355] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.934772] env[62066]: DEBUG oslo_concurrency.lockutils [req-6291897d-c3c9-49d1-9072-5bad3ed8f4d8 req-2e63e2c2-7910-47d9-ab83-0854f015e572 service nova] Releasing lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.038295] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d2b24b-1295-6f39-51db-9862b9bcda1b, 'name': SearchDatastore_Task, 'duration_secs': 0.009946} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.039084] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-347a28fc-4d68-4df1-a190-e04ed2fda63e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.045175] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 836.045175] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5295f2c7-1bd5-bc14-c9ef-b39534807ec5" [ 836.045175] env[62066]: _type = "Task" [ 836.045175] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.054398] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5295f2c7-1bd5-bc14-c9ef-b39534807ec5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.159546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.052s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.162405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.162649] env[62066]: DEBUG nova.objects.instance [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lazy-loading 'resources' on Instance uuid 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.195485] env[62066]: INFO nova.scheduler.client.report [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Deleted allocations for instance 4723be94-3479-4e66-8088-914824c0e669 [ 836.296284] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156208, 'name': Rename_Task, 'duration_secs': 0.180602} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.296583] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.296835] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e183473f-64e6-4e08-89d5-987e77e28ec5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.305435] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 836.305435] env[62066]: value = "task-1156210" [ 836.305435] env[62066]: _type = "Task" [ 836.305435] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.314091] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.322301] env[62066]: DEBUG oslo_vmware.api [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156209, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20188} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.322553] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.322742] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.322924] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.323119] env[62066]: INFO nova.compute.manager [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 836.323368] env[62066]: DEBUG oslo.service.loopingcall [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.323551] env[62066]: DEBUG nova.compute.manager [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.323647] env[62066]: DEBUG nova.network.neutron [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 836.447531] env[62066]: INFO nova.compute.manager [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Took 27.87 seconds to build instance. [ 836.470030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.564303] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5295f2c7-1bd5-bc14-c9ef-b39534807ec5, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.564303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.564303] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.564303] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3957fb73-229e-4eff-829a-cf3a40942ebb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.572118] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 836.572118] env[62066]: value = "task-1156211" [ 836.572118] env[62066]: _type = "Task" [ 836.572118] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.581483] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.696569] env[62066]: DEBUG nova.compute.manager [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Stashing vm_state: stopped {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 836.704068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fc67501-cd37-4222-8bef-a9f8f2ef18a4 tempest-ServerRescueTestJSONUnderV235-1924915666 tempest-ServerRescueTestJSONUnderV235-1924915666-project-member] Lock "4723be94-3479-4e66-8088-914824c0e669" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.168s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.821904] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156210, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.950287] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c70fb05-1f94-423a-b5eb-96d5d03ee086 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.383s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.088582] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156211, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.091015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a95ae-6302-413c-a0aa-42ebb66db4f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.102229] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430c86c2-13e3-43e2-b3c3-732aae8af376 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.139831] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308ea5ee-4109-4b88-8418-e083913f7325 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.149857] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438ea7fd-73e4-4ef5-b61b-2b2003c4e4da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.166478] env[62066]: DEBUG nova.network.neutron [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.168117] env[62066]: DEBUG nova.compute.provider_tree [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.209866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.210149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.210339] env[62066]: DEBUG nova.compute.manager [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.211283] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ead1bd2-233e-49d1-9898-9b28dc11b617 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.218934] env[62066]: DEBUG nova.compute.manager [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 837.220205] env[62066]: DEBUG nova.objects.instance [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'flavor' on Instance uuid 68bf0e89-8236-44ef-b514-d25698eb23c3 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.225044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.317064] env[62066]: DEBUG oslo_vmware.api [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156210, 'name': PowerOnVM_Task, 'duration_secs': 0.517717} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.317897] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.317897] env[62066]: INFO nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Took 8.99 seconds to spawn the instance on the hypervisor. [ 837.317897] env[62066]: DEBUG nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.318559] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4acabb-ee49-481f-a760-e792ebe8983e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.586387] env[62066]: DEBUG nova.compute.manager [req-05c2bb1f-5e8b-4886-b8ba-311a67a1d5fa req-fb74c4c3-3dfa-49c6-bb2a-33f7ad4df60a service nova] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Received event network-vif-deleted-6383bfb0-3350-4704-bc27-2450027edc30 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.591343] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514578} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.591965] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.592466] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.593639] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdb5c394-dd0a-421b-ac7d-4c82dcaee5db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.602947] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 837.602947] env[62066]: value = "task-1156212" [ 837.602947] env[62066]: _type = "Task" [ 837.602947] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.612936] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156212, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.671227] env[62066]: INFO nova.compute.manager [-] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Took 1.35 seconds to deallocate network for instance. [ 837.672408] env[62066]: DEBUG nova.scheduler.client.report [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.724147] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 837.724531] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53de91f2-42f5-4d04-af27-5c20cb591745 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.732245] env[62066]: DEBUG oslo_vmware.api [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 837.732245] env[62066]: value = "task-1156213" [ 837.732245] env[62066]: _type = "Task" [ 837.732245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.741041] env[62066]: DEBUG oslo_vmware.api [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.839817] env[62066]: INFO nova.compute.manager [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Took 27.97 seconds to build instance. [ 838.121023] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156212, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075525} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.121023] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.121023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fe4e8c-8fae-4056-a13f-737cdd864b40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.150881] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.151182] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f21220ef-b3f0-41a6-a524-47d7c1de71b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.182172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.020s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.185927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.800s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.187690] env[62066]: INFO nova.compute.claims [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.191245] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.191649] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 838.191649] env[62066]: value = "task-1156214" [ 838.191649] env[62066]: _type = "Task" [ 838.191649] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.202584] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.222941] env[62066]: INFO nova.scheduler.client.report [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Deleted allocations for instance 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf [ 838.248029] env[62066]: DEBUG oslo_vmware.api [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156213, 'name': PowerOffVM_Task, 'duration_secs': 0.392107} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.248171] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 838.248409] env[62066]: DEBUG nova.compute.manager [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.250202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03fe308-5067-46dd-ba83-0b4fe8a1a524 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.342147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0a9ebe94-1d71-4b80-bc46-5ed3ef6d1196 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.485s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.710326] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156214, 'name': ReconfigVM_Task, 'duration_secs': 0.322543} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.710326] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfigured VM instance instance-0000004a to attach disk [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.710326] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-059150b6-f7d3-4c82-852b-a8c24f7d3252 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.717875] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 838.717875] env[62066]: value = "task-1156215" [ 838.717875] env[62066]: _type = "Task" [ 838.717875] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.730567] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156215, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.731110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d8afd27-8613-4b9d-8d4d-4c5b33a000f1 tempest-ListImageFiltersTestJSON-2101403784 tempest-ListImageFiltersTestJSON-2101403784-project-member] Lock "01e1df17-4b9d-4e12-bf6b-50b39c08bfbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.535s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.764160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45b7e027-c4d1-4cac-924b-d986e58a4556 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.152191] env[62066]: DEBUG nova.compute.manager [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Received event network-changed-f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.152453] env[62066]: DEBUG nova.compute.manager [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Refreshing instance network info cache due to event network-changed-f19a58d1-80e6-4f51-bd80-36927b0c3867. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.153436] env[62066]: DEBUG oslo_concurrency.lockutils [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] Acquiring lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.153436] env[62066]: DEBUG oslo_concurrency.lockutils [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] Acquired lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.153436] env[62066]: DEBUG nova.network.neutron [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Refreshing network info cache for port f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.232311] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156215, 'name': Rename_Task, 'duration_secs': 0.156132} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.233016] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.233343] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bd22aae-e186-47ac-bb7b-df9b28e4acfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.243052] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 839.243052] env[62066]: value = "task-1156216" [ 839.243052] env[62066]: _type = "Task" [ 839.243052] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.256384] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.601232] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6fab3a-e700-43b7-9a66-9c5b7e279773 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.611973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1918bcc2-59fa-45d7-91ce-fb091d475a39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.673202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7c1973-7692-41c2-b672-dd53ec00bdb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.690026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b9d484-1c45-4b0f-8569-25669ee89d73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.710987] env[62066]: DEBUG nova.compute.provider_tree [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.754017] env[62066]: DEBUG oslo_vmware.api [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156216, 'name': PowerOnVM_Task, 'duration_secs': 0.502928} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.754485] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.754769] env[62066]: INFO nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Took 7.83 seconds to spawn the instance on the hypervisor. [ 839.755010] env[62066]: DEBUG nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.755949] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e1eb7e-ab55-4b0a-9dd0-9ab957a370a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.092263] env[62066]: DEBUG nova.network.neutron [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updated VIF entry in instance network info cache for port f19a58d1-80e6-4f51-bd80-36927b0c3867. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.092667] env[62066]: DEBUG nova.network.neutron [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updating instance_info_cache with network_info: [{"id": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "address": "fa:16:3e:59:33:1f", "network": {"id": "fd494499-759b-4d73-a332-60a6f9a126af", "bridge": "br-int", "label": "tempest-ServersTestJSON-207731434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25052e3fae44617b5a63f881a7b081b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf19a58d1-80", "ovs_interfaceid": "f19a58d1-80e6-4f51-bd80-36927b0c3867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.218024] env[62066]: DEBUG nova.scheduler.client.report [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.273489] env[62066]: INFO nova.compute.manager [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Took 18.59 seconds to build instance. [ 840.415514] env[62066]: INFO nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Rebuilding instance [ 840.466336] env[62066]: DEBUG nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.467238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc64f88-07f8-467c-85b9-0e47ff9908c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.595771] env[62066]: DEBUG oslo_concurrency.lockutils [req-b27450ad-6d19-4ab9-b3fe-f800689858d7 req-922fc8da-5c48-4914-a939-26c9eadbb499 service nova] Releasing lock "refresh_cache-3092f591-214f-40ba-a8e2-ccc6a0f007c9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.723713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.724550] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.732018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.882s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.732018] env[62066]: DEBUG nova.objects.instance [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lazy-loading 'pci_requests' on Instance uuid a2251e56-2787-412d-89c9-eef111ee6d2b {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.776240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6a76dce1-a596-47c7-b6fe-650e7c7eb5af tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.710s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.985659] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.985659] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d538c98-067d-42d2-a7c2-1ff6062ec46a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.994020] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 840.994020] env[62066]: value = "task-1156217" [ 840.994020] env[62066]: _type = "Task" [ 840.994020] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.006352] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 841.006352] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.006352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c63313-db2f-4b84-9d5e-c1e59e1c205b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.014967] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.015365] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-825e3056-8ab9-43ea-8ed0-0915d1049a97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.083792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.084067] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.084264] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore1] 68bf0e89-8236-44ef-b514-d25698eb23c3 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.084780] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6dbe0236-0890-419b-ab65-fa2afa97b837 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.099314] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 841.099314] env[62066]: value = "task-1156219" [ 841.099314] env[62066]: _type = "Task" [ 841.099314] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.111855] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.233796] env[62066]: DEBUG nova.compute.utils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 841.239213] env[62066]: DEBUG nova.objects.instance [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lazy-loading 'numa_topology' on Instance uuid a2251e56-2787-412d-89c9-eef111ee6d2b {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.239213] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.239213] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 841.343078] env[62066]: DEBUG nova.policy [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.613614] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25935} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.613614] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.613614] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.613947] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.712838] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Successfully created port: 63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.739809] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.743151] env[62066]: INFO nova.compute.claims [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.296796] env[62066]: DEBUG nova.compute.manager [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.668150] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.668629] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.668629] env[62066]: DEBUG nova.virt.hardware [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.669347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5555dab0-7431-4469-9cfa-5e86f061b380 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.681275] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2613f70f-748a-4c05-bbcf-fb6758a69cfe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.693382] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce30c59-f9f3-4578-923a-d5d1962866f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.706537] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:14:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94c39e5f-27a8-4373-9893-8defce990708', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.714245] env[62066]: DEBUG oslo.service.loopingcall [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.715128] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.715440] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc91f962-9f5a-46dd-9f24-960689f5912b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.734670] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8283373-c7f6-4ec9-b99f-804b3bccc981 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.740221] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.740221] env[62066]: value = "task-1156220" [ 842.740221] env[62066]: _type = "Task" [ 842.740221] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.780975] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.782879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05341768-0642-43b6-82b9-7b28b3165b55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.789253] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156220, 'name': CreateVM_Task} progress is 15%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.795631] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caa0edf-acbd-4fd4-ac1b-58273c70b811 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.820696] env[62066]: DEBUG nova.compute.provider_tree [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.823121] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.823276] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.823486] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.823759] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.823919] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.824129] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.824929] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.824929] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.824929] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.824929] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.825125] env[62066]: DEBUG nova.virt.hardware [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.825960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "0fda5f59-55ac-4150-8402-00064d14c8ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.825960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.829193] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b36d7f-3fbf-40b6-bf28-c2f81fe46439 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.832666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.841044] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef903284-c786-4cf6-aaee-f9d324c08263 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.861933] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "f578eda3-3d81-418b-bcb7-0a954835ed72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.862247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.898760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.898760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.257686] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156220, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.335060] env[62066]: DEBUG nova.scheduler.client.report [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.339042] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.364636] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.400385] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.470419] env[62066]: DEBUG nova.compute.manager [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Received event network-vif-plugged-63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.470665] env[62066]: DEBUG oslo_concurrency.lockutils [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] Acquiring lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.470911] env[62066]: DEBUG oslo_concurrency.lockutils [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.471246] env[62066]: DEBUG oslo_concurrency.lockutils [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.471246] env[62066]: DEBUG nova.compute.manager [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] No waiting events found dispatching network-vif-plugged-63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.471830] env[62066]: WARNING nova.compute.manager [req-bd3841cc-3a62-433d-b9e4-f7f84c680eab req-9398d774-c53e-49fe-8b73-1faba34fa04a service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Received unexpected event network-vif-plugged-63b01da8-ba06-4b05-893b-3278a175b040 for instance with vm_state building and task_state spawning. [ 843.570094] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Successfully updated port: 63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.760103] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156220, 'name': CreateVM_Task, 'duration_secs': 0.626718} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.760103] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.760103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.760103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.760103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.760103] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c23bc686-1933-4136-82a4-c4704bd042d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.766439] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 843.766439] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523da546-1dbf-51f2-f96d-b7a112993f54" [ 843.766439] env[62066]: _type = "Task" [ 843.766439] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.775943] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523da546-1dbf-51f2-f96d-b7a112993f54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.844575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.116s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.849485] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.073s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.851188] env[62066]: INFO nova.compute.claims [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.873422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.890432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.903140] env[62066]: INFO nova.network.neutron [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating port 4991ab48-c8b7-437b-81d7-db93c7d3da4d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 843.928474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.074685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.074877] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.075103] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.281807] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523da546-1dbf-51f2-f96d-b7a112993f54, 'name': SearchDatastore_Task, 'duration_secs': 0.041474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.282690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.283626] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.283982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.284251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.285669] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.285669] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9096e93-1a4b-41f1-aac9-c31b68627ee1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.300017] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.300017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.300017] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f29913a4-88a7-4218-adb1-a41b1b1d52e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.306293] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 844.306293] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527a48a2-0fb4-1d18-c5ad-d373da44beaf" [ 844.306293] env[62066]: _type = "Task" [ 844.306293] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.319669] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a48a2-0fb4-1d18-c5ad-d373da44beaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.606812] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.754634] env[62066]: DEBUG nova.network.neutron [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Updating instance_info_cache with network_info: [{"id": "63b01da8-ba06-4b05-893b-3278a175b040", "address": "fa:16:3e:9d:e0:15", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b01da8-ba", "ovs_interfaceid": "63b01da8-ba06-4b05-893b-3278a175b040", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.821106] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527a48a2-0fb4-1d18-c5ad-d373da44beaf, 'name': SearchDatastore_Task, 'duration_secs': 0.012537} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.822623] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-813cec67-284f-41da-ad3d-72fd335deced {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.829766] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 844.829766] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd0a4-d4c4-0235-c61f-9ce9dcad0dab" [ 844.829766] env[62066]: _type = "Task" [ 844.829766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.839820] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd0a4-d4c4-0235-c61f-9ce9dcad0dab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.257561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.257937] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Instance network_info: |[{"id": "63b01da8-ba06-4b05-893b-3278a175b040", "address": "fa:16:3e:9d:e0:15", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b01da8-ba", "ovs_interfaceid": "63b01da8-ba06-4b05-893b-3278a175b040", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.258449] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:e0:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63b01da8-ba06-4b05-893b-3278a175b040', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.267379] env[62066]: DEBUG oslo.service.loopingcall [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.270109] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.270609] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b015d869-56a5-47b8-8a0d-18c68ec45ecf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.288877] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cf236d-6471-4f49-a3c9-fb0bf363462c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.298583] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f3150a-78d4-4a17-9bd6-1c076cc8c765 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.301704] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.301704] env[62066]: value = "task-1156221" [ 845.301704] env[62066]: _type = "Task" [ 845.301704] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.340751] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a952e4-c455-479f-a267-746e1149cf64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.346696] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156221, 'name': CreateVM_Task} progress is 15%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.360023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa97fe9d-f7c6-4ab2-aa65-ab142f930682 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.360931] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ccd0a4-d4c4-0235-c61f-9ce9dcad0dab, 'name': SearchDatastore_Task, 'duration_secs': 0.010145} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.361244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.361537] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.362248] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b0d682f-d13a-4f57-be06-c5aa134dfeff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.375261] env[62066]: DEBUG nova.compute.provider_tree [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.382501] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 845.382501] env[62066]: value = "task-1156222" [ 845.382501] env[62066]: _type = "Task" [ 845.382501] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.392660] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156222, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.549534] env[62066]: DEBUG nova.compute.manager [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Received event network-changed-63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.549773] env[62066]: DEBUG nova.compute.manager [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Refreshing instance network info cache due to event network-changed-63b01da8-ba06-4b05-893b-3278a175b040. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.550065] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Acquiring lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.550237] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Acquired lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.550414] env[62066]: DEBUG nova.network.neutron [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Refreshing network info cache for port 63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.560323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.560521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.560750] env[62066]: DEBUG nova.network.neutron [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.650575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.652895] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.813240] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156221, 'name': CreateVM_Task, 'duration_secs': 0.361727} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.813470] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.814232] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.814413] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.814776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.815179] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5a0df5c-ef71-4ce3-a90c-97b5b394e1b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.822665] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 845.822665] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526b6a79-c869-7648-d018-dd533370db61" [ 845.822665] env[62066]: _type = "Task" [ 845.822665] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.835555] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526b6a79-c869-7648-d018-dd533370db61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.881019] env[62066]: DEBUG nova.scheduler.client.report [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.894486] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156222, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.157900] env[62066]: DEBUG nova.compute.utils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 846.334744] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526b6a79-c869-7648-d018-dd533370db61, 'name': SearchDatastore_Task, 'duration_secs': 0.058495} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.337634] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.337956] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.338266] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.338479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.338729] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.339092] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dff1df5-55ad-4985-b4c5-fed4701ea8ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.362963] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.363344] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.364218] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2ff6d3c-4c2a-48fa-b5ca-004c307bf001 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.370379] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 846.370379] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52546d82-7e1b-8cf0-a658-3d06763fc3be" [ 846.370379] env[62066]: _type = "Task" [ 846.370379] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.382694] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52546d82-7e1b-8cf0-a658-3d06763fc3be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.384727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.386176] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.389104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.850s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.393155] env[62066]: INFO nova.compute.claims [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.406775] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156222, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724416} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.407790] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.408151] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.408463] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc0a36a7-95fd-4a64-8fdf-0705f08ce414 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.416685] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 846.416685] env[62066]: value = "task-1156223" [ 846.416685] env[62066]: _type = "Task" [ 846.416685] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.427875] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.468117] env[62066]: DEBUG nova.network.neutron [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Updated VIF entry in instance network info cache for port 63b01da8-ba06-4b05-893b-3278a175b040. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.468540] env[62066]: DEBUG nova.network.neutron [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Updating instance_info_cache with network_info: [{"id": "63b01da8-ba06-4b05-893b-3278a175b040", "address": "fa:16:3e:9d:e0:15", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63b01da8-ba", "ovs_interfaceid": "63b01da8-ba06-4b05-893b-3278a175b040", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.485013] env[62066]: DEBUG nova.network.neutron [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.660209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.882435] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52546d82-7e1b-8cf0-a658-3d06763fc3be, 'name': SearchDatastore_Task, 'duration_secs': 0.046152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.883253] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6beefcd5-ef09-4df6-9e96-4e8ff892f7ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.888709] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 846.888709] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52487b62-2ed2-d49d-4116-4c2dfcc8a109" [ 846.888709] env[62066]: _type = "Task" [ 846.888709] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.896539] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52487b62-2ed2-d49d-4116-4c2dfcc8a109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.899168] env[62066]: DEBUG nova.compute.utils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 846.904021] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.904021] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 846.926670] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072502} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.927034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.927899] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32340a31-c220-4a8b-8c5f-cfc949c14e7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.952662] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.955229] env[62066]: DEBUG nova.policy [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e55565e67934c319f216673b2e35479', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8788711396624222b7ab712e4906ead3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.957060] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1482b27-17c6-4860-b98f-f68a07504a46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.973430] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Releasing lock "refresh_cache-6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.973775] env[62066]: DEBUG nova.compute.manager [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.974050] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.974320] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.974547] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.974777] env[62066]: DEBUG nova.compute.manager [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] No waiting events found dispatching network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.975032] env[62066]: WARNING nova.compute.manager [req-f2ec2530-4b13-4d49-88af-904c4d31940a req-eafa5c0e-72ee-45dc-a5b2-70125cb3b3fe service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received unexpected event network-vif-plugged-4991ab48-c8b7-437b-81d7-db93c7d3da4d for instance with vm_state shelved_offloaded and task_state spawning. [ 846.982015] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 846.982015] env[62066]: value = "task-1156224" [ 846.982015] env[62066]: _type = "Task" [ 846.982015] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.987284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.993375] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156224, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c4875149733e09f4d082a50ab69761a6',container_format='bare',created_at=2024-10-07T00:57:09Z,direct_url=,disk_format='vmdk',id=ca07540e-1efc-45fb-a554-db31a70bc000,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1603224318-shelved',owner='7ab8a3b607934456b5d408442f03524a',properties=ImageMetaProps,protected=,size=31661056,status='active',tags=,updated_at=2024-10-07T00:57:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.019335] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.019716] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.019811] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.020033] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.020292] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.020529] env[62066]: DEBUG nova.virt.hardware [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.021443] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdcd69f-a7ac-4d84-b1d0-213b5dfc5af9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.032135] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbbe1cb-18ba-4fe4-91cf-c7f5eb2b9920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.050213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:29:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7b4bfde-f109-4f64-adab-e7f06b80685d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4991ab48-c8b7-437b-81d7-db93c7d3da4d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.059429] env[62066]: DEBUG oslo.service.loopingcall [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.060052] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 847.060327] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70d68ab5-f9f2-4587-9216-2f5f0d26d83c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.082939] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.082939] env[62066]: value = "task-1156225" [ 847.082939] env[62066]: _type = "Task" [ 847.082939] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.092701] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156225, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.330933] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Successfully created port: 3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.402545] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52487b62-2ed2-d49d-4116-4c2dfcc8a109, 'name': SearchDatastore_Task, 'duration_secs': 0.029804} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.403413] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.413904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.414318] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15/6bd9da25-fcfc-41a1-b5d6-07d28ab37e15.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.418114] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a61e6b1f-96ab-4b7c-ac83-0ae4fdba9e36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.429333] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 847.429333] env[62066]: value = "task-1156226" [ 847.429333] env[62066]: _type = "Task" [ 847.429333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.439969] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156226, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.493073] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.595947] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156225, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.762049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.762406] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.762610] env[62066]: INFO nova.compute.manager [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Attaching volume c902f613-58dc-4ab2-9b93-fe59b6b4cee2 to /dev/sdb [ 847.816188] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcfb8f7-2630-49b2-b7ec-463edc0dba7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.821492] env[62066]: DEBUG nova.compute.manager [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.821691] env[62066]: DEBUG nova.compute.manager [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing instance network info cache due to event network-changed-4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.821912] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] Acquiring lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.822104] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] Acquired lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.822292] env[62066]: DEBUG nova.network.neutron [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Refreshing network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.831890] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3876ad49-3a33-4585-8752-11fa5853c9d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.852467] env[62066]: DEBUG nova.virt.block_device [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating existing volume attachment record: c8008450-053d-4f97-8883-865f998082cb {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 847.907871] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041f36b6-284f-4e85-b906-210efef7b794 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.918428] env[62066]: INFO nova.virt.block_device [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Booting with volume 1f4e0b34-91c4-413f-8225-3151a6e125ab at /dev/sda [ 847.922541] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a482222-5f16-4f4f-a507-97c52e8dd262 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.972609] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e734605-5726-483c-9cfe-f18f4ad044f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.973989] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07d4641d-b205-41f0-8a67-fca87eb8511f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.980528] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156226, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.991606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3749394b-de2b-4cbd-af1a-38ef925b6530 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.998998] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a958f029-ad44-4aa1-a894-e4c8963f2d36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.012900] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156224, 'name': ReconfigVM_Task, 'duration_secs': 0.609514} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.013839] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 68bf0e89-8236-44ef-b514-d25698eb23c3/68bf0e89-8236-44ef-b514-d25698eb23c3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.014601] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01533567-29db-4f96-a047-e58e1f587b95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.024908] env[62066]: DEBUG nova.compute.provider_tree [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.038230] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 848.038230] env[62066]: value = "task-1156227" [ 848.038230] env[62066]: _type = "Task" [ 848.038230] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.039384] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab5af03b-fffa-4cc9-91c5-f7305b56622c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.056619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af356565-c943-4fb6-9801-8d41d6349e07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.067341] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156227, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.094853] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a1cc84-8880-4113-bde9-a27a4c9ef3f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.103380] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156225, 'name': CreateVM_Task, 'duration_secs': 0.612065} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.105405] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 848.106245] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.106422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.106835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.107619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3a2005-26ba-4b28-b0db-a0e23efe2103 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.110202] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-588d93f2-1ebd-4395-8fec-05072eb620f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.115394] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 848.115394] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b5eaf2-08c6-7539-2375-51c9d61f3802" [ 848.115394] env[62066]: _type = "Task" [ 848.115394] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.124431] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b5eaf2-08c6-7539-2375-51c9d61f3802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.128099] env[62066]: DEBUG nova.virt.block_device [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updating existing volume attachment record: 157bcb6b-288c-4515-abf5-7751b19e0c0c {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 848.444684] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156226, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658107} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.444684] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15/6bd9da25-fcfc-41a1-b5d6-07d28ab37e15.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.444684] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.444684] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9fcdf8e8-4cb8-45b5-b095-8eda7dfa0645 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.452548] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 848.452548] env[62066]: value = "task-1156231" [ 848.452548] env[62066]: _type = "Task" [ 848.452548] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.463101] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.540449] env[62066]: DEBUG nova.scheduler.client.report [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.553435] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156227, 'name': Rename_Task, 'duration_secs': 0.239816} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.553766] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.554051] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-031a48e6-4fa4-4a50-8cbf-c6416d7cefc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.561523] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 848.561523] env[62066]: value = "task-1156232" [ 848.561523] env[62066]: _type = "Task" [ 848.561523] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.565185] env[62066]: DEBUG nova.network.neutron [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updated VIF entry in instance network info cache for port 4991ab48-c8b7-437b-81d7-db93c7d3da4d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.565525] env[62066]: DEBUG nova.network.neutron [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [{"id": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "address": "fa:16:3e:d9:29:d3", "network": {"id": "00c898eb-fb4d-4f95-8910-f8ed293bc183", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1223295728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab8a3b607934456b5d408442f03524a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4991ab48-c8", "ovs_interfaceid": "4991ab48-c8b7-437b-81d7-db93c7d3da4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.572258] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156232, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.627270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.627536] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Processing image ca07540e-1efc-45fb-a554-db31a70bc000 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.627774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.627924] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.628129] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.628504] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6a8653-a79d-4470-abdc-2c9fc248590c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.644978] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.645202] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.645960] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c02fdf68-bf50-4bd1-b81f-4c0397b6f811 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.652025] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 848.652025] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d3df7e-7bbf-c466-3b97-82def678d7ec" [ 848.652025] env[62066]: _type = "Task" [ 848.652025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.660816] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d3df7e-7bbf-c466-3b97-82def678d7ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.941849] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Successfully updated port: 3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.965457] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081296} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.965457] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.965457] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a500eaeb-b9c7-428a-bc8b-865e19f04bfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.988941] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15/6bd9da25-fcfc-41a1-b5d6-07d28ab37e15.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.989935] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c551c75-04f9-4364-a734-1a31dff0f073 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.010582] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 849.010582] env[62066]: value = "task-1156233" [ 849.010582] env[62066]: _type = "Task" [ 849.010582] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.019287] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.048760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.659s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.048760] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.051702] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.623s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.053286] env[62066]: INFO nova.compute.claims [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.068220] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b41d43d-7e13-4fa9-83aa-a63167fb31e2 req-704f4f77-0b27-4de1-aae0-5325f74735a5 service nova] Releasing lock "refresh_cache-a2251e56-2787-412d-89c9-eef111ee6d2b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.073419] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156232, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.166657] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 849.167156] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Fetch image to [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55/OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 849.167508] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Downloading stream optimized image ca07540e-1efc-45fb-a554-db31a70bc000 to [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55/OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55.vmdk on the data store datastore2 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 849.167769] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Downloading image file data ca07540e-1efc-45fb-a554-db31a70bc000 to the ESX as VM named 'OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 849.259054] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 849.259054] env[62066]: value = "resgroup-9" [ 849.259054] env[62066]: _type = "ResourcePool" [ 849.259054] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 849.259407] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-068fc66a-b279-4d3b-a8cc-0cf144eb030e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.283434] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lease: (returnval){ [ 849.283434] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 849.283434] env[62066]: _type = "HttpNfcLease" [ 849.283434] env[62066]: } obtained for vApp import into resource pool (val){ [ 849.283434] env[62066]: value = "resgroup-9" [ 849.283434] env[62066]: _type = "ResourcePool" [ 849.283434] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 849.283787] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the lease: (returnval){ [ 849.283787] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 849.283787] env[62066]: _type = "HttpNfcLease" [ 849.283787] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 849.291270] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.291270] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 849.291270] env[62066]: _type = "HttpNfcLease" [ 849.291270] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 849.444757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.444757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquired lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.444974] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.523173] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156233, 'name': ReconfigVM_Task, 'duration_secs': 0.372039} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.523485] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15/6bd9da25-fcfc-41a1-b5d6-07d28ab37e15.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.524090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8497e37-e429-483a-adbc-11222e270173 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.531120] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 849.531120] env[62066]: value = "task-1156235" [ 849.531120] env[62066]: _type = "Task" [ 849.531120] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.540117] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156235, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.558302] env[62066]: DEBUG nova.compute.utils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 849.562928] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.563224] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.583648] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156232, 'name': PowerOnVM_Task, 'duration_secs': 0.807052} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.584063] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.584309] env[62066]: DEBUG nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.585280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff0c4db-bd80-4082-8e2a-2086d90775d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.610429] env[62066]: DEBUG nova.policy [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.793878] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.793878] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 849.793878] env[62066]: _type = "HttpNfcLease" [ 849.793878] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 849.847809] env[62066]: DEBUG nova.compute.manager [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Received event network-vif-plugged-3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.848056] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Acquiring lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.848278] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.848453] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.848624] env[62066]: DEBUG nova.compute.manager [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] No waiting events found dispatching network-vif-plugged-3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.848792] env[62066]: WARNING nova.compute.manager [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Received unexpected event network-vif-plugged-3a5c1815-3b73-48c3-af45-080c8bbb02dc for instance with vm_state building and task_state spawning. [ 849.848954] env[62066]: DEBUG nova.compute.manager [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Received event network-changed-3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.849158] env[62066]: DEBUG nova.compute.manager [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Refreshing instance network info cache due to event network-changed-3a5c1815-3b73-48c3-af45-080c8bbb02dc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.849339] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Acquiring lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.904212] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Successfully created port: 0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.986671] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.041610] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156235, 'name': Rename_Task, 'duration_secs': 0.164184} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.041939] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.042253] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d99295c1-34d6-4f1b-9cc5-79a69228fe23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.050356] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 850.050356] env[62066]: value = "task-1156236" [ 850.050356] env[62066]: _type = "Task" [ 850.050356] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.058641] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.063266] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.098982] env[62066]: INFO nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] bringing vm to original state: 'stopped' [ 850.223731] env[62066]: DEBUG nova.network.neutron [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updating instance_info_cache with network_info: [{"id": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "address": "fa:16:3e:60:4f:aa", "network": {"id": "11d3a4f1-f061-4ecf-acc1-0e6c607a5f1d", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-828561973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8788711396624222b7ab712e4906ead3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a5c1815-3b", "ovs_interfaceid": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.231676] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.232371] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.232702] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.232943] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.233199] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.233373] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.237230] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.237515] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.237658] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.238266] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.238266] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.238266] env[62066]: DEBUG nova.virt.hardware [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.242501] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af9aca1-7beb-4042-85b6-66cc370f8d9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.252170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e43e8b1-47bf-48d4-89e5-120b822f2e23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.294668] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.294668] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 850.294668] env[62066]: _type = "HttpNfcLease" [ 850.294668] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 850.480933] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5f9af9-9315-4a3f-88b4-9c7c665cdbc7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.489216] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73581a04-4413-4dfb-8ed9-b554f7f44026 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.520711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e54e63c-a108-414f-ba01-a2767070e606 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.530230] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3703bc7d-fa27-4e13-917b-1fc191bc9f99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.545149] env[62066]: DEBUG nova.compute.provider_tree [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.561628] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156236, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.726895] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Releasing lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.727282] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance network_info: |[{"id": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "address": "fa:16:3e:60:4f:aa", "network": {"id": "11d3a4f1-f061-4ecf-acc1-0e6c607a5f1d", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-828561973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8788711396624222b7ab712e4906ead3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a5c1815-3b", "ovs_interfaceid": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.727618] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Acquired lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.727858] env[62066]: DEBUG nova.network.neutron [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Refreshing network info cache for port 3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.729206] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:4f:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3a80436-f7a9-431a-acec-aca3d76e3f9b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a5c1815-3b73-48c3-af45-080c8bbb02dc', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.736866] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Creating folder: Project (8788711396624222b7ab712e4906ead3). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.737979] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b71fbf8a-4df7-4302-9e59-71e443aa7baf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.752105] env[62066]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 850.752365] env[62066]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62066) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 850.752725] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Folder already exists: Project (8788711396624222b7ab712e4906ead3). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 850.752964] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Creating folder: Instances. Parent ref: group-v251683. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.753255] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ae1bd77-b5b0-435d-9cfd-f363c0752cac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.763771] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Created folder: Instances in parent group-v251683. [ 850.764020] env[62066]: DEBUG oslo.service.loopingcall [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.764223] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.764435] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8209370-45ed-4ad3-927c-c2c4ba8531d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.785834] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.785834] env[62066]: value = "task-1156240" [ 850.785834] env[62066]: _type = "Task" [ 850.785834] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.798645] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156240, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.800192] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.800192] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 850.800192] env[62066]: _type = "HttpNfcLease" [ 850.800192] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 851.048840] env[62066]: DEBUG nova.scheduler.client.report [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.063184] env[62066]: DEBUG oslo_vmware.api [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156236, 'name': PowerOnVM_Task, 'duration_secs': 0.533147} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.063510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.063828] env[62066]: INFO nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Took 8.28 seconds to spawn the instance on the hypervisor. [ 851.063909] env[62066]: DEBUG nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.065383] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8b87df-7f1c-4f86-a821-c76dd4b293de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.077260] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.107596] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.107933] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.108139] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.108388] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.108526] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.108830] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.108979] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.109197] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.109439] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.109727] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.109977] env[62066]: DEBUG nova.virt.hardware [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.110953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.110953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.110953] env[62066]: DEBUG nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.112175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7cc729-654c-46b5-8ce9-fc7229aa7dad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.116149] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144f55d1-c760-4959-8d93-1b5b3ddf2120 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.128060] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfa0e63-8ab0-4a2f-ba01-e3da1872b971 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.133580] env[62066]: DEBUG nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 851.138079] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.138422] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e0b7aa0-4f5a-49aa-a4d7-d42b03f58ade {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.156542] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 851.156542] env[62066]: value = "task-1156241" [ 851.156542] env[62066]: _type = "Task" [ 851.156542] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.168594] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.298125] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.298125] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 851.298125] env[62066]: _type = "HttpNfcLease" [ 851.298125] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 851.301708] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 851.301708] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e7c445-cb38-6090-c429-11f1a61fa1c3" [ 851.301708] env[62066]: _type = "HttpNfcLease" [ 851.301708] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 851.301925] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156240, 'name': CreateVM_Task, 'duration_secs': 0.447015} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.302650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d32aaf-fbb5-4744-a421-22769474d8cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.305150] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.305855] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'device_type': None, 'boot_index': 0, 'attachment_id': '157bcb6b-288c-4515-abf5-7751b19e0c0c', 'mount_device': '/dev/sda', 'guest_format': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251692', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'name': 'volume-1f4e0b34-91c4-413f-8225-3151a6e125ab', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753f585a-22d7-4eeb-8580-4e3a68b5fd72', 'attached_at': '', 'detached_at': '', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'serial': '1f4e0b34-91c4-413f-8225-3151a6e125ab'}, 'volume_type': None}], 'swap': None} {{(pid=62066) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 851.306082] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Root volume attach. Driver type: vmdk {{(pid=62066) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 851.307208] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3eef06-f014-4546-87bc-02050d34bfc2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.313008] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 851.313197] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating HTTP connection to write to file with size = 31661056 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 851.381481] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70cf1d0-091a-41cc-ac42-865e7222ed4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.390105] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2c23cbf9-1565-4129-af07-eb17c07b9e45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.394167] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5d7e9e-a6c2-4afb-8b3d-0624d5affd03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.403285] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e1384bc5-a5a2-453c-879f-36932aaea0e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.412049] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 851.412049] env[62066]: value = "task-1156242" [ 851.412049] env[62066]: _type = "Task" [ 851.412049] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.421220] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156242, 'name': RelocateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.557332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.557872] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.561833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.122s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.563356] env[62066]: INFO nova.compute.claims [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.592226] env[62066]: INFO nova.compute.manager [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Took 26.23 seconds to build instance. [ 851.684484] env[62066]: DEBUG oslo_vmware.api [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156241, 'name': PowerOffVM_Task, 'duration_secs': 0.185088} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.684484] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.684808] env[62066]: DEBUG nova.compute.manager [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.689504] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b0c179-acbd-4530-aace-75ba461cc8ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.821214] env[62066]: DEBUG nova.network.neutron [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updated VIF entry in instance network info cache for port 3a5c1815-3b73-48c3-af45-080c8bbb02dc. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.821465] env[62066]: DEBUG nova.network.neutron [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updating instance_info_cache with network_info: [{"id": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "address": "fa:16:3e:60:4f:aa", "network": {"id": "11d3a4f1-f061-4ecf-acc1-0e6c607a5f1d", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-828561973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8788711396624222b7ab712e4906ead3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a5c1815-3b", "ovs_interfaceid": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.902236] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Successfully updated port: 0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.930762] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156242, 'name': RelocateVM_Task, 'duration_secs': 0.455507} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.932734] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 851.933020] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251692', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'name': 'volume-1f4e0b34-91c4-413f-8225-3151a6e125ab', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753f585a-22d7-4eeb-8580-4e3a68b5fd72', 'attached_at': '', 'detached_at': '', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'serial': '1f4e0b34-91c4-413f-8225-3151a6e125ab'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 851.933913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd66815e-da36-4397-af3d-f52cc061950f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.957839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db30c5b-1745-4bc7-8e9f-60ce03bb8d61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.982296] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-1f4e0b34-91c4-413f-8225-3151a6e125ab/volume-1f4e0b34-91c4-413f-8225-3151a6e125ab.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.986267] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-324ed60c-c478-41db-ab35-a5149d720ac5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.014459] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 852.014459] env[62066]: value = "task-1156243" [ 852.014459] env[62066]: _type = "Task" [ 852.014459] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.016659] env[62066]: DEBUG nova.compute.manager [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-plugged-0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.016882] env[62066]: DEBUG oslo_concurrency.lockutils [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.017130] env[62066]: DEBUG oslo_concurrency.lockutils [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.017313] env[62066]: DEBUG oslo_concurrency.lockutils [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.017576] env[62066]: DEBUG nova.compute.manager [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] No waiting events found dispatching network-vif-plugged-0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.017704] env[62066]: WARNING nova.compute.manager [req-b9dc86f3-8dd1-47a6-bd84-1887a7914ea3 req-c9466f89-7aab-4712-9da1-df4090e22a9f service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received unexpected event network-vif-plugged-0d8007bd-9e20-4780-a21e-a22c8c7dac13 for instance with vm_state building and task_state spawning. [ 852.032841] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156243, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.068082] env[62066]: DEBUG nova.compute.utils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 852.071103] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.072020] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.094215] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3367a34-dd62-406f-9504-c68978934e79 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.753s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.114457] env[62066]: DEBUG nova.policy [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0757382f267042ecbc68f0c08607ed93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a5a550637cd748b9ae9988f2cf838c2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.218109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.107s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.325333] env[62066]: DEBUG oslo_concurrency.lockutils [req-e67d2aa3-f2b5-48de-9a9e-bd29e4092a6a req-3b6b5aba-a1f4-4ae8-952c-01300c96f079 service nova] Releasing lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.339033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.339033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.339033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.339033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.339033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.343809] env[62066]: INFO nova.compute.manager [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Terminating instance [ 852.348409] env[62066]: DEBUG nova.compute.manager [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.349097] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.350442] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98560409-6a38-4b0c-8131-612f63688d99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.368583] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.369434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-668b7bf4-ef0f-4ca5-b24a-d8f7fb2244c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.382141] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 852.382141] env[62066]: value = "task-1156244" [ 852.382141] env[62066]: _type = "Task" [ 852.382141] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.397577] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.407686] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.407940] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.408043] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.434562] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 852.434910] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251705', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'name': 'volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '831c0bed-8a41-4672-aa08-d60b64f365b8', 'attached_at': '', 'detached_at': '', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'serial': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 852.435822] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1312e19f-7af4-476f-bc61-b300f1197db3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.460736] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc98d32-73a7-408a-b40f-07812f5c68e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.489834] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2/volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.491283] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Successfully created port: 6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.497949] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05bcb05b-e79a-45c5-8eea-9c5a2aea6e0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.526098] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 852.526098] env[62066]: value = "task-1156245" [ 852.526098] env[62066]: _type = "Task" [ 852.526098] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.534744] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156243, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.541731] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.578424] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.585326] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 852.585406] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 852.586761] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36b4435-5578-4556-8026-fffa4b7b3716 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.595156] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 852.595376] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 852.595735] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a5a640d8-ffd9-4d06-8d36-19820cf72495 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.730165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.862796] env[62066]: DEBUG oslo_vmware.rw_handles [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b36ee3-5421-b855-0356-729bd5a29619/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 852.863065] env[62066]: INFO nova.virt.vmwareapi.images [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Downloaded image file data ca07540e-1efc-45fb-a554-db31a70bc000 [ 852.864071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe67c50-fd34-4239-8348-abadc19929c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.891906] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-893e4cde-18cf-488f-b62f-9ea615fbb6e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.899745] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156244, 'name': PowerOffVM_Task, 'duration_secs': 0.262962} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.900043] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.900235] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.900480] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-536d6e6c-8cec-4e69-8e74-ec4382209543 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.918825] env[62066]: INFO nova.virt.vmwareapi.images [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] The imported VM was unregistered [ 852.921708] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 852.921975] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Creating directory with path [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.925520] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f44986cb-9cce-4301-b8c9-4da9b30d86c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.938727] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Created directory with path [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.939010] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55/OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55.vmdk to [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 852.947822] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-853e0a7d-4bdb-4178-9570-5b6ae123c34d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.949899] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.959231] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 852.959231] env[62066]: value = "task-1156248" [ 852.959231] env[62066]: _type = "Task" [ 852.959231] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.973654] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.985237] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.985464] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.985884] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.986186] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-444f3231-24dc-4763-8414-861ba1e2a1b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.994630] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 852.994630] env[62066]: value = "task-1156249" [ 852.994630] env[62066]: _type = "Task" [ 852.994630] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.008678] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156249, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.029630] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156243, 'name': ReconfigVM_Task, 'duration_secs': 0.518789} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.038727] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-1f4e0b34-91c4-413f-8225-3151a6e125ab/volume-1f4e0b34-91c4-413f-8225-3151a6e125ab.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.044037] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-251f9b7e-07e7-4d13-99fc-2b072a1e32af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.064666] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156245, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.066367] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 853.066367] env[62066]: value = "task-1156250" [ 853.066367] env[62066]: _type = "Task" [ 853.066367] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.075240] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156250, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.096556] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244f14ec-aaef-4ba4-adb4-6ae9ded04397 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.107865] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2b848e-a8df-46ae-a514-28a10d82888c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.145098] env[62066]: DEBUG nova.network.neutron [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.147069] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e907021a-346b-4053-8e8d-dfe6c7f341d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.163014] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1327436e-7033-4ae2-ad62-69e6bc99c389 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.184030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.184030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.184030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.184030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.184030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.186074] env[62066]: DEBUG nova.compute.provider_tree [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.188769] env[62066]: INFO nova.compute.manager [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Terminating instance [ 853.193025] env[62066]: DEBUG nova.compute.manager [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.193025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.193238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e04e6c1-a1e3-478e-a5f6-807cfecd58a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.207907] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.208272] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dffa07e0-37ec-42df-9f12-aae7ffa51c0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.336017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.336017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.336017] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore2] 68bf0e89-8236-44ef-b514-d25698eb23c3 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.336243] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23930e10-d5b5-44bd-ab75-1d120e27de56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.346283] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 853.346283] env[62066]: value = "task-1156252" [ 853.346283] env[62066]: _type = "Task" [ 853.346283] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.356910] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.472206] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.509134] env[62066]: DEBUG oslo_vmware.api [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156249, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273297} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.509498] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.509728] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.510054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.510128] env[62066]: INFO nova.compute.manager [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Took 1.16 seconds to destroy the instance on the hypervisor. [ 853.510378] env[62066]: DEBUG oslo.service.loopingcall [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.510576] env[62066]: DEBUG nova.compute.manager [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.510667] env[62066]: DEBUG nova.network.neutron [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 853.541987] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156245, 'name': ReconfigVM_Task, 'duration_secs': 0.545278} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.542532] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2/volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.548306] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4df26d9-c031-4e6f-8507-c4f34d684de0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.567573] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 853.567573] env[62066]: value = "task-1156253" [ 853.567573] env[62066]: _type = "Task" [ 853.567573] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.586536] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156250, 'name': ReconfigVM_Task, 'duration_secs': 0.148153} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.586766] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156253, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.587093] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251692', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'name': 'volume-1f4e0b34-91c4-413f-8225-3151a6e125ab', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753f585a-22d7-4eeb-8580-4e3a68b5fd72', 'attached_at': '', 'detached_at': '', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'serial': '1f4e0b34-91c4-413f-8225-3151a6e125ab'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 853.587690] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d837552a-cc89-4109-ad9a-232118470132 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.598480] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 853.598480] env[62066]: value = "task-1156254" [ 853.598480] env[62066]: _type = "Task" [ 853.598480] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.602925] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.613073] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156254, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.634188] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.634451] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.635040] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.635040] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.635040] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.635232] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.635308] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.635523] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.635713] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.635889] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.636138] env[62066]: DEBUG nova.virt.hardware [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.637086] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61a2fbc-48f3-4be6-8d4d-a9a9380f8937 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.650047] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e21ffb-327e-4e4d-9b18-4577fcfe2ee7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.655119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.655515] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Instance network_info: |[{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.655981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:60:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '65497291-07f3-434c-bd42-657a0cb03365', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0d8007bd-9e20-4780-a21e-a22c8c7dac13', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.664172] env[62066]: DEBUG oslo.service.loopingcall [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.664957] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.665172] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea48bffd-d6fc-4546-9fbd-20999c5d1d4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.690228] env[62066]: DEBUG nova.scheduler.client.report [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.696611] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.696611] env[62066]: value = "task-1156255" [ 853.696611] env[62066]: _type = "Task" [ 853.696611] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.709453] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156255, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.859432] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.977618] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.084196] env[62066]: DEBUG oslo_vmware.api [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156253, 'name': ReconfigVM_Task, 'duration_secs': 0.175802} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.084561] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251705', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'name': 'volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '831c0bed-8a41-4672-aa08-d60b64f365b8', 'attached_at': '', 'detached_at': '', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'serial': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 854.103840] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Successfully updated port: 6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.114403] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156254, 'name': Rename_Task, 'duration_secs': 0.167224} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.114666] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.115247] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a03c7e8-2dc5-4d64-9d76-750e7be16c07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.127691] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 854.127691] env[62066]: value = "task-1156256" [ 854.127691] env[62066]: _type = "Task" [ 854.127691] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.139793] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156256, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.148121] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-changed-0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.148377] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing instance network info cache due to event network-changed-0d8007bd-9e20-4780-a21e-a22c8c7dac13. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.148561] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.148728] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.148863] env[62066]: DEBUG nova.network.neutron [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing network info cache for port 0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.196036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.196485] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.199762] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.354s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.199940] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.200172] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 854.200857] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.349s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.200857] env[62066]: DEBUG nova.objects.instance [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lazy-loading 'resources' on Instance uuid 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.202858] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d46193-403f-4112-9005-a86784dec6d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.217153] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156255, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.221132] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d43b67c-a603-4d70-97d6-d136f9fce19d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.240152] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1bc980-e764-4b31-ac05-439bfc696466 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.251254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d76b7e-4dfc-4fc1-9776-6be5a000c7ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.288418] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179774MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 854.288601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.298290] env[62066]: DEBUG nova.network.neutron [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.358902] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.473465] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.607828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.610371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.610801] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.651082] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156256, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.704970] env[62066]: DEBUG nova.compute.utils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 854.714563] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.714788] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.731551] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156255, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.801499] env[62066]: DEBUG nova.policy [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fa978afbf3541588bca1af1045442b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a912137933243bab1057a6e80fda023', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.803401] env[62066]: INFO nova.compute.manager [-] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Took 1.29 seconds to deallocate network for instance. [ 854.863130] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.978025] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.017102] env[62066]: DEBUG nova.network.neutron [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updated VIF entry in instance network info cache for port 0d8007bd-9e20-4780-a21e-a22c8c7dac13. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 855.017102] env[62066]: DEBUG nova.network.neutron [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.130193] env[62066]: DEBUG nova.objects.instance [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'flavor' on Instance uuid 831c0bed-8a41-4672-aa08-d60b64f365b8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.164812] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156256, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.166109] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.217661] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156255, 'name': CreateVM_Task, 'duration_secs': 1.417859} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.221383] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.222169] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.227054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.227195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.227559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.228360] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-414a0f30-ffb3-4160-b172-f36e2c74ca8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.242554] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 855.242554] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ded0-49a6-11a2-c17e-6668a9eea12a" [ 855.242554] env[62066]: _type = "Task" [ 855.242554] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.242554] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ddf482-4570-4b32-9c3b-cd002d126877 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.255398] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ded0-49a6-11a2-c17e-6668a9eea12a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.260141] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68dd1b1e-78bc-4938-a6ee-e0115a78523c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.266369] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Successfully created port: ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.304124] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33076cd-d85e-458a-9cf5-a26a01d053d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.313385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.317614] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d82bcff-e8ef-4a59-9b74-a4cae503a9dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.340022] env[62066]: DEBUG nova.compute.provider_tree [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.365315] env[62066]: DEBUG oslo_vmware.api [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.850272} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.365315] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.365315] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.365315] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.365315] env[62066]: INFO nova.compute.manager [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Took 2.17 seconds to destroy the instance on the hypervisor. [ 855.365315] env[62066]: DEBUG oslo.service.loopingcall [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.365315] env[62066]: DEBUG nova.compute.manager [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.365315] env[62066]: DEBUG nova.network.neutron [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.445419] env[62066]: DEBUG nova.network.neutron [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.477726] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.520652] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.521100] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Received event network-vif-deleted-63b01da8-ba06-4b05-893b-3278a175b040 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.521213] env[62066]: INFO nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Neutron deleted interface 63b01da8-ba06-4b05-893b-3278a175b040; detaching it from the instance and deleting it from the info cache [ 855.521416] env[62066]: DEBUG nova.network.neutron [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.641834] env[62066]: DEBUG oslo_vmware.api [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156256, 'name': PowerOnVM_Task, 'duration_secs': 1.401157} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.642224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.642434] env[62066]: INFO nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Took 5.41 seconds to spawn the instance on the hypervisor. [ 855.642625] env[62066]: DEBUG nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.643478] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54915e9-05ff-42a0-bdcc-512b6f2b408a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.648648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02f30f72-891a-4812-8c6e-0b66f163a23a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.886s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.649937] env[62066]: INFO nova.compute.manager [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Rescuing [ 855.650167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.650167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.650556] env[62066]: DEBUG nova.network.neutron [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.755820] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b9ded0-49a6-11a2-c17e-6668a9eea12a, 'name': SearchDatastore_Task, 'duration_secs': 0.085646} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.756223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.756562] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.756866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.757093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.757344] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.757658] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29ee61aa-0719-4582-bba6-4ce58dc9d9d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.766215] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.766436] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.768333] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3a017e2-af58-4a18-8468-b0c14e0b2fb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.775741] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 855.775741] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5277f6be-11f2-3a57-ea35-303e79f6b14c" [ 855.775741] env[62066]: _type = "Task" [ 855.775741] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.786016] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5277f6be-11f2-3a57-ea35-303e79f6b14c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.841831] env[62066]: DEBUG nova.scheduler.client.report [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.873863] env[62066]: DEBUG nova.compute.manager [req-127eefd7-1a34-402a-a89b-0c6fcf8d9947 req-090477ef-074f-4811-ac74-cae6552a9e89 service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Received event network-vif-deleted-94c39e5f-27a8-4373-9893-8defce990708 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.873863] env[62066]: INFO nova.compute.manager [req-127eefd7-1a34-402a-a89b-0c6fcf8d9947 req-090477ef-074f-4811-ac74-cae6552a9e89 service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Neutron deleted interface 94c39e5f-27a8-4373-9893-8defce990708; detaching it from the instance and deleting it from the info cache [ 855.873863] env[62066]: DEBUG nova.network.neutron [req-127eefd7-1a34-402a-a89b-0c6fcf8d9947 req-090477ef-074f-4811-ac74-cae6552a9e89 service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.949032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.949032] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Instance network_info: |[{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.949624] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:23:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6244eb3b-c073-494b-b711-f8712351da75', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.956993] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Creating folder: Project (a5a550637cd748b9ae9988f2cf838c2f). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.957583] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7fcbe67-ba01-4fe1-a284-34a2aa09d72f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.969019] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Created folder: Project (a5a550637cd748b9ae9988f2cf838c2f) in parent group-v251573. [ 855.969217] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Creating folder: Instances. Parent ref: group-v251710. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.969429] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8de2b2d1-1ba4-4fd1-b6e2-4f2e35a11277 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.974243] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156248, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.518741} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.974780] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55/OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55.vmdk to [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk. [ 855.974974] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Cleaning up location [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 855.975165] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_349b51c9-2bda-49a8-94e3-fe0560dd6e55 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.975389] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-240f0743-0a13-450d-a5c7-881d1ef89510 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.982214] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 855.982214] env[62066]: value = "task-1156259" [ 855.982214] env[62066]: _type = "Task" [ 855.982214] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.986258] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Created folder: Instances in parent group-v251710. [ 855.986485] env[62066]: DEBUG oslo.service.loopingcall [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.986955] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.987176] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2494a8c-ccfc-4ad8-89fd-dbd964accf48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.004031] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156259, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.008311] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.008311] env[62066]: value = "task-1156260" [ 856.008311] env[62066]: _type = "Task" [ 856.008311] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.016826] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156260, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.023862] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a1c94a7-44e3-4b1c-a4d7-a2e51cd4b048 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.031214] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c62605-9451-4e52-8e70-eeaee40f6b93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.061499] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Detach interface failed, port_id=63b01da8-ba06-4b05-893b-3278a175b040, reason: Instance 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 856.062116] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Received event network-vif-plugged-6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.062116] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.062229] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.062320] env[62066]: DEBUG oslo_concurrency.lockutils [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.062494] env[62066]: DEBUG nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] No waiting events found dispatching network-vif-plugged-6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.062667] env[62066]: WARNING nova.compute.manager [req-93ecd7c7-c2a9-48ed-8cd0-be20e0b39a0e req-7bd42a81-0b0f-49dc-83b6-74d317d1bf7d service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Received unexpected event network-vif-plugged-6244eb3b-c073-494b-b711-f8712351da75 for instance with vm_state building and task_state spawning. [ 856.168682] env[62066]: INFO nova.compute.manager [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Took 27.41 seconds to build instance. [ 856.199457] env[62066]: DEBUG nova.compute.manager [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Received event network-changed-6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.199681] env[62066]: DEBUG nova.compute.manager [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Refreshing instance network info cache due to event network-changed-6244eb3b-c073-494b-b711-f8712351da75. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.199900] env[62066]: DEBUG oslo_concurrency.lockutils [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] Acquiring lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.200057] env[62066]: DEBUG oslo_concurrency.lockutils [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] Acquired lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.200223] env[62066]: DEBUG nova.network.neutron [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Refreshing network info cache for port 6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.236505] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.261214] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.261504] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.261670] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.261859] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.262015] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.262173] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.262408] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.262582] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.262753] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.262917] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.263108] env[62066]: DEBUG nova.virt.hardware [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.264016] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effe19b9-8529-436b-979f-8c20fbc7385e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.273360] env[62066]: DEBUG nova.network.neutron [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.277295] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bf1574-ffae-4563-b13b-3e36456fc33f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.299038] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5277f6be-11f2-3a57-ea35-303e79f6b14c, 'name': SearchDatastore_Task, 'duration_secs': 0.009831} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.300089] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41daba84-5734-4d7b-9ded-f256a915f4d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.305824] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 856.305824] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524a0f17-e2c8-e1de-5348-7bd8ca294efe" [ 856.305824] env[62066]: _type = "Task" [ 856.305824] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.314478] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524a0f17-e2c8-e1de-5348-7bd8ca294efe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.347077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.146s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.349440] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.036s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.351380] env[62066]: INFO nova.compute.claims [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.376486] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bafb621d-5475-45f2-b0e8-4fe892bdb60d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.387155] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970d132b-fb4b-4df4-bdba-73bce7f0fb82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.403331] env[62066]: INFO nova.scheduler.client.report [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted allocations for instance 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4 [ 856.427831] env[62066]: DEBUG nova.compute.manager [req-127eefd7-1a34-402a-a89b-0c6fcf8d9947 req-090477ef-074f-4811-ac74-cae6552a9e89 service nova] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Detach interface failed, port_id=94c39e5f-27a8-4373-9893-8defce990708, reason: Instance 68bf0e89-8236-44ef-b514-d25698eb23c3 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 856.492756] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156259, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036578} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.493864] env[62066]: DEBUG nova.network.neutron [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.495063] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.495258] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.495553] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk to [datastore2] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.496019] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1818c3d-b2d2-4495-aeb3-ec7778f37741 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.502871] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 856.502871] env[62066]: value = "task-1156261" [ 856.502871] env[62066]: _type = "Task" [ 856.502871] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.511710] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.519807] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156260, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.670952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-439bd576-6dc7-4098-9c55-6673f2c448b9 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.939s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.780929] env[62066]: INFO nova.compute.manager [-] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Took 1.42 seconds to deallocate network for instance. [ 856.821694] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524a0f17-e2c8-e1de-5348-7bd8ca294efe, 'name': SearchDatastore_Task, 'duration_secs': 0.010468} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.822072] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.822393] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f24f8067-07b2-4941-8464-c30b58983ec1/f24f8067-07b2-4941-8464-c30b58983ec1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.822842] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d178ae48-8301-40fc-a0fd-80f9ec8e577a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.835119] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 856.835119] env[62066]: value = "task-1156262" [ 856.835119] env[62066]: _type = "Task" [ 856.835119] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.848591] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.899928] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Successfully updated port: ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.912595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3db5b08-3493-4a13-9bc5-f3f49bda6b01 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.235s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.997289] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.024128] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.038730] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156260, 'name': CreateVM_Task, 'duration_secs': 0.727315} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.038730] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.040015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.040015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.040015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.040371] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8c778eb-e0ec-479b-af11-70523fd7cd31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.047267] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 857.047267] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52840f38-e913-ea36-ae71-50753d598172" [ 857.047267] env[62066]: _type = "Task" [ 857.047267] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.063873] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52840f38-e913-ea36-ae71-50753d598172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.169324] env[62066]: DEBUG nova.network.neutron [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updated VIF entry in instance network info cache for port 6244eb3b-c073-494b-b711-f8712351da75. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.169898] env[62066]: DEBUG nova.network.neutron [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.288872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.348158] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.401913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.402209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquired lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.402611] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.519314] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.543250] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.543812] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c25ddb6-ee60-4c11-8efe-3099a338e40d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.556677] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 857.556677] env[62066]: value = "task-1156263" [ 857.556677] env[62066]: _type = "Task" [ 857.556677] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.568599] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52840f38-e913-ea36-ae71-50753d598172, 'name': SearchDatastore_Task, 'duration_secs': 0.08647} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.572427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.572661] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.572909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.573100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.573296] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.573643] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.573901] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a68e00b3-ced8-4c81-9a53-6704f19abd6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.596410] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.596640] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.597501] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5de78c-b70d-4526-90f8-07e85c861793 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.606200] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 857.606200] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52835bc8-015d-e9a8-275d-92b98ac7d2fd" [ 857.606200] env[62066]: _type = "Task" [ 857.606200] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.620192] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52835bc8-015d-e9a8-275d-92b98ac7d2fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.674836] env[62066]: DEBUG oslo_concurrency.lockutils [req-cbe26bdc-c6ba-4d9d-9818-8c970e4ba782 req-0f78935d-52d0-4e99-a145-3320c66cbf21 service nova] Releasing lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.759255] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bc00be-db2e-4cad-8ce0-ce64d266ffb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.772495] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4259bb-c23c-4957-872c-38153234cb34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.810960] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e858ba3-8ebf-4ff1-ba51-e017b4a6fe2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.821798] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83759499-9446-4fdd-8ea9-87562644bf19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.842200] env[62066]: DEBUG nova.compute.provider_tree [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.856175] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.961244] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.017074] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.075264] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156263, 'name': PowerOffVM_Task, 'duration_secs': 0.218907} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.075776] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.076941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98501be0-1ab2-46c6-adf3-13f9bfb3cb99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.115839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cbbeea-594a-47cd-b21c-6034d6d06220 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.132498] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52835bc8-015d-e9a8-275d-92b98ac7d2fd, 'name': SearchDatastore_Task, 'duration_secs': 0.091785} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.132498] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c877307b-58dd-42ed-b7c4-195b5ff7ae64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.140824] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 858.140824] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52fc7dd7-1be1-6ae3-1239-959f7f633d3b" [ 858.140824] env[62066]: _type = "Task" [ 858.140824] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.154631] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fc7dd7-1be1-6ae3-1239-959f7f633d3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.160179] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.160179] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-566e17aa-aeea-43c6-acd3-5139caeb149c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.169265] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 858.169265] env[62066]: value = "task-1156264" [ 858.169265] env[62066]: _type = "Task" [ 858.169265] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.185020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 858.185020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.185020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.348070] env[62066]: DEBUG nova.scheduler.client.report [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.358227] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.421490] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Received event network-changed-3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.421770] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Refreshing instance network info cache due to event network-changed-3a5c1815-3b73-48c3-af45-080c8bbb02dc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.421976] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Acquiring lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.422145] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Acquired lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.422352] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Refreshing network info cache for port 3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.454027] env[62066]: DEBUG nova.network.neutron [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Updating instance_info_cache with network_info: [{"id": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "address": "fa:16:3e:a0:5f:28", "network": {"id": "1a35249f-2bd1-4454-948e-d382b230a3fa", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-414823095-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a912137933243bab1057a6e80fda023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0e826d-c9", "ovs_interfaceid": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.518472] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.656083] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fc7dd7-1be1-6ae3-1239-959f7f633d3b, 'name': SearchDatastore_Task, 'duration_secs': 0.085716} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.656245] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.656524] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 31e50f97-f873-44ad-9923-67923cdb8d3a/31e50f97-f873-44ad-9923-67923cdb8d3a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.656836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.657049] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.657289] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59f58b0a-b85e-4715-ae53-b86720df4def {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.659979] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d208e352-cf0b-43fb-a235-ee4b6b49f8b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.669084] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 858.669084] env[62066]: value = "task-1156265" [ 858.669084] env[62066]: _type = "Task" [ 858.669084] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.680556] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.682056] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.682187] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.682859] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74f07a00-ff35-4ffc-a944-9d00faf7b6b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.691415] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 858.691415] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52fe0cf9-3983-2b12-d244-8da873811c65" [ 858.691415] env[62066]: _type = "Task" [ 858.691415] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.702331] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fe0cf9-3983-2b12-d244-8da873811c65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.855859] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.859925] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.860495] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.863355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.393s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.864953] env[62066]: INFO nova.compute.claims [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.956569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Releasing lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.956945] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Instance network_info: |[{"id": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "address": "fa:16:3e:a0:5f:28", "network": {"id": "1a35249f-2bd1-4454-948e-d382b230a3fa", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-414823095-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a912137933243bab1057a6e80fda023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0e826d-c9", "ovs_interfaceid": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.957451] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:5f:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d0c6fd7-3cc9-4818-9475-8f15900394cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec0e826d-c94d-4295-a202-c1e2bfdcefd5', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.969980] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Creating folder: Project (7a912137933243bab1057a6e80fda023). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.970361] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d30c2b03-5d14-4031-a8aa-2c5953f46118 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.985891] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Created folder: Project (7a912137933243bab1057a6e80fda023) in parent group-v251573. [ 858.986128] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Creating folder: Instances. Parent ref: group-v251713. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.986392] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b44e5f1-7c77-4911-a6b0-997c309ca476 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.996218] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Created folder: Instances in parent group-v251713. [ 858.996490] env[62066]: DEBUG oslo.service.loopingcall [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.996722] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.996968] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c5c7763-66d3-4214-8a72-b8977331d3e4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.024731] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.024731] env[62066]: value = "task-1156268" [ 859.024731] env[62066]: _type = "Task" [ 859.024731] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.028393] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.039339] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156268, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.179935] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.203781] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fe0cf9-3983-2b12-d244-8da873811c65, 'name': SearchDatastore_Task, 'duration_secs': 0.078541} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.207045] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-add8a483-766a-44df-94b5-111f8066ada0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.213078] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 859.213078] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d7477c-186b-fe42-2550-b482e9caaf23" [ 859.213078] env[62066]: _type = "Task" [ 859.213078] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.222058] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d7477c-186b-fe42-2550-b482e9caaf23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.354924] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156262, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.377727} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.355263] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f24f8067-07b2-4941-8464-c30b58983ec1/f24f8067-07b2-4941-8464-c30b58983ec1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.355490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.355756] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6e079ea-a82a-4a42-90b8-a0daf2ddb122 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.364470] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 859.364470] env[62066]: value = "task-1156269" [ 859.364470] env[62066]: _type = "Task" [ 859.364470] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.369888] env[62066]: DEBUG nova.compute.utils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 859.372190] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.372285] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 859.375929] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updated VIF entry in instance network info cache for port 3a5c1815-3b73-48c3-af45-080c8bbb02dc. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.376362] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updating instance_info_cache with network_info: [{"id": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "address": "fa:16:3e:60:4f:aa", "network": {"id": "11d3a4f1-f061-4ecf-acc1-0e6c607a5f1d", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-828561973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8788711396624222b7ab712e4906ead3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3a80436-f7a9-431a-acec-aca3d76e3f9b", "external-id": "cl2-zone-339", "segmentation_id": 339, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a5c1815-3b", "ovs_interfaceid": "3a5c1815-3b73-48c3-af45-080c8bbb02dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.383795] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156269, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.445083] env[62066]: DEBUG nova.policy [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.532029] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156261, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.542593} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.535540] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ca07540e-1efc-45fb-a554-db31a70bc000/ca07540e-1efc-45fb-a554-db31a70bc000.vmdk to [datastore2] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.536451] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8ff969-9b00-4959-9e5b-a393eca5759b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.561856] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.565681] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e69c488f-8082-4228-8579-0e4e7e34e4f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.580834] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156268, 'name': CreateVM_Task, 'duration_secs': 0.51004} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.581061] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.582298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.582479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.582830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.583130] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ed69641-167e-467f-8e7b-abe620d318c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.590269] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 859.590269] env[62066]: value = "task-1156270" [ 859.590269] env[62066]: _type = "Task" [ 859.590269] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.591910] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 859.591910] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52406f0a-e758-7d97-3779-a6e7b0c87271" [ 859.591910] env[62066]: _type = "Task" [ 859.591910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.604692] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156270, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.608069] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52406f0a-e758-7d97-3779-a6e7b0c87271, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.680071] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156265, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.726998] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d7477c-186b-fe42-2550-b482e9caaf23, 'name': SearchDatastore_Task, 'duration_secs': 0.018253} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.726998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.726998] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 859.726998] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df6d555e-5125-4143-9177-f7a8d64d61eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.736208] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 859.736208] env[62066]: value = "task-1156271" [ 859.736208] env[62066]: _type = "Task" [ 859.736208] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.744754] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.764311] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Successfully created port: 13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.879871] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.882171] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156269, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093647} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.882474] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.883365] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcb776c-9503-4d20-ab54-7e48a460dc37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.888172] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Releasing lock "refresh_cache-753f585a-22d7-4eeb-8580-4e3a68b5fd72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.888449] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Received event network-vif-plugged-ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.888657] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Acquiring lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.889189] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.889440] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.889676] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] No waiting events found dispatching network-vif-plugged-ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.889897] env[62066]: WARNING nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Received unexpected event network-vif-plugged-ec0e826d-c94d-4295-a202-c1e2bfdcefd5 for instance with vm_state building and task_state spawning. [ 859.890178] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Received event network-changed-ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.890372] env[62066]: DEBUG nova.compute.manager [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Refreshing instance network info cache due to event network-changed-ec0e826d-c94d-4295-a202-c1e2bfdcefd5. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.892018] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Acquiring lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.892018] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Acquired lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.892018] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Refreshing network info cache for port ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.929690] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] f24f8067-07b2-4941-8464-c30b58983ec1/f24f8067-07b2-4941-8464-c30b58983ec1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.934976] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca2c10e6-669e-4f2e-a147-ffdd906d4566 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.963491] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 859.963491] env[62066]: value = "task-1156272" [ 859.963491] env[62066]: _type = "Task" [ 859.963491] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.978274] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156272, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.106900] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156270, 'name': ReconfigVM_Task, 'duration_secs': 0.363703} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.110931] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Reconfigured VM instance instance-00000037 to attach disk [datastore2] a2251e56-2787-412d-89c9-eef111ee6d2b/a2251e56-2787-412d-89c9-eef111ee6d2b.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.111781] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52406f0a-e758-7d97-3779-a6e7b0c87271, 'name': SearchDatastore_Task, 'duration_secs': 0.054952} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.115589] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abd7e75f-bd55-46db-a1cb-a7299dd96028 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.116530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.116778] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.117023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.117176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.117358] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.117923] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0727a14-daad-478f-8d3a-5d8c87da9d20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.127185] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 860.127185] env[62066]: value = "task-1156273" [ 860.127185] env[62066]: _type = "Task" [ 860.127185] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.136038] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.136257] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.137103] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e20097d-bbe2-4a24-be81-cbc14f3244ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.142869] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156273, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.149341] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 860.149341] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527723a5-6fbe-2945-a356-9052aa3ab2ed" [ 860.149341] env[62066]: _type = "Task" [ 860.149341] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.161762] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527723a5-6fbe-2945-a356-9052aa3ab2ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.185057] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156265, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.020514} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.185480] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 31e50f97-f873-44ad-9923-67923cdb8d3a/31e50f97-f873-44ad-9923-67923cdb8d3a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.185731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.186164] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00a6740d-edcb-4b21-8465-4e4edb1d45ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.198321] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 860.198321] env[62066]: value = "task-1156274" [ 860.198321] env[62066]: _type = "Task" [ 860.198321] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.211208] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.251331] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156271, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.347724] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd8d353-82ec-41f1-b959-1226ea6c017d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.355833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee71d3b3-70c8-41b3-b222-d4eae5499e64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.388302] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316f4c72-0b43-4b13-a990-2d8079b08f46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.400483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d34f57-7b13-45ba-a114-10a7a47d1116 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.418127] env[62066]: DEBUG nova.compute.provider_tree [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.474745] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156272, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.640578] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156273, 'name': Rename_Task, 'duration_secs': 0.209685} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.640978] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.641299] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-667cf6e9-ad14-4e82-a183-22182f2254d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.648137] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 860.648137] env[62066]: value = "task-1156275" [ 860.648137] env[62066]: _type = "Task" [ 860.648137] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.666506] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527723a5-6fbe-2945-a356-9052aa3ab2ed, 'name': SearchDatastore_Task, 'duration_secs': 0.05304} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.666860] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.667734] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af522e85-9012-456a-8313-d3827f8f3e4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.674115] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 860.674115] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dde816-6ac4-f292-da87-7736c0083a1b" [ 860.674115] env[62066]: _type = "Task" [ 860.674115] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.685526] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52dde816-6ac4-f292-da87-7736c0083a1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.708492] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156274, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115234} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.708492] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.709326] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eeb1a6-c91e-42ff-a905-3e727027c6f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.712720] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Updated VIF entry in instance network info cache for port ec0e826d-c94d-4295-a202-c1e2bfdcefd5. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.713122] env[62066]: DEBUG nova.network.neutron [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Updating instance_info_cache with network_info: [{"id": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "address": "fa:16:3e:a0:5f:28", "network": {"id": "1a35249f-2bd1-4454-948e-d382b230a3fa", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-414823095-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a912137933243bab1057a6e80fda023", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec0e826d-c9", "ovs_interfaceid": "ec0e826d-c94d-4295-a202-c1e2bfdcefd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.736283] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 31e50f97-f873-44ad-9923-67923cdb8d3a/31e50f97-f873-44ad-9923-67923cdb8d3a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.737343] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd22ea5f-cb7f-49ba-801f-fc1bf5fd701f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.761643] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531803} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.762901] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk. [ 860.763272] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 860.763272] env[62066]: value = "task-1156276" [ 860.763272] env[62066]: _type = "Task" [ 860.763272] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.763975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f03c53-acd1-4539-b0a9-d2b68fb3009c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.775748] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.801404] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.801774] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46734f9d-5efc-4502-a066-99146b49fe4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.821751] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 860.821751] env[62066]: value = "task-1156277" [ 860.821751] env[62066]: _type = "Task" [ 860.821751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.831116] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156277, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.896269] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.920758] env[62066]: DEBUG nova.scheduler.client.report [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.932314] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.932580] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.932770] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.932963] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.933195] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.933381] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.933619] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.933805] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.934042] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.934260] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.934446] env[62066]: DEBUG nova.virt.hardware [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.935417] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321ceb7c-9e35-4ad9-9e56-7e36f5c21919 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.947266] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e927dc64-6a7f-4c96-a1c2-3af34bd53c94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.976353] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156272, 'name': ReconfigVM_Task, 'duration_secs': 1.009327} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.977409] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfigured VM instance instance-0000004d to attach disk [datastore2] f24f8067-07b2-4941-8464-c30b58983ec1/f24f8067-07b2-4941-8464-c30b58983ec1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.978150] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b59f9847-4a75-4030-8e12-936a89f50f42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.987813] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 860.987813] env[62066]: value = "task-1156278" [ 860.987813] env[62066]: _type = "Task" [ 860.987813] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.997105] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156278, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.159745] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156275, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.186766] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52dde816-6ac4-f292-da87-7736c0083a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.01108} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.186766] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.187107] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da8b3bca-dbf5-4266-826d-ed93772fa04c/da8b3bca-dbf5-4266-826d-ed93772fa04c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.187244] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c07e447e-a036-45aa-ac9c-1fa967dfae18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.195783] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 861.195783] env[62066]: value = "task-1156279" [ 861.195783] env[62066]: _type = "Task" [ 861.195783] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.205393] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.216592] env[62066]: DEBUG oslo_concurrency.lockutils [req-2f9a8c38-d7f0-4071-98bc-5b1e51377604 req-7f9d90ea-3a70-4369-a6ee-c7b561b54754 service nova] Releasing lock "refresh_cache-da8b3bca-dbf5-4266-826d-ed93772fa04c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.278760] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156276, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.327614] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Successfully updated port: 13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.336806] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156277, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.358436] env[62066]: DEBUG nova.compute.manager [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Received event network-vif-plugged-13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.358778] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] Acquiring lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.359098] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.359293] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.359495] env[62066]: DEBUG nova.compute.manager [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] No waiting events found dispatching network-vif-plugged-13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.359692] env[62066]: WARNING nova.compute.manager [req-a4d19a20-1d97-4529-8f50-a0ccea1e55f8 req-5063d291-294f-42cb-976a-51587105511f service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Received unexpected event network-vif-plugged-13c8bbfc-3aad-4881-966e-45eca3162138 for instance with vm_state building and task_state spawning. [ 861.427780] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.428030] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.432656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 24.208s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.500216] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156278, 'name': Rename_Task, 'duration_secs': 0.405771} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.500523] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.502558] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b207869-9467-43ed-b5ff-069600528626 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.508583] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 861.508583] env[62066]: value = "task-1156280" [ 861.508583] env[62066]: _type = "Task" [ 861.508583] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.520900] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156280, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.659226] env[62066]: DEBUG oslo_vmware.api [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156275, 'name': PowerOnVM_Task, 'duration_secs': 0.525802} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.659831] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.706486] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156279, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.767762] env[62066]: DEBUG nova.compute.manager [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.768797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e7a329-e011-443c-bb8a-3592081b4fb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.781155] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156276, 'name': ReconfigVM_Task, 'duration_secs': 0.836578} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.783214] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 31e50f97-f873-44ad-9923-67923cdb8d3a/31e50f97-f873-44ad-9923-67923cdb8d3a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.786762] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-886aaeba-4b26-4016-baf3-2753f0b7d33d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.796810] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 861.796810] env[62066]: value = "task-1156281" [ 861.796810] env[62066]: _type = "Task" [ 861.796810] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.809377] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156281, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.830163] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.830394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.830394] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.835612] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156277, 'name': ReconfigVM_Task, 'duration_secs': 0.777163} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.836217] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8/fc5145ed-66bc-4490-b8ac-7ca0de814dc0-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.837202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53433d8-cd8a-4482-929f-1c6143580579 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.874796] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bd695df-7a49-4c55-9245-2231a6c0b8ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.895487] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 861.895487] env[62066]: value = "task-1156282" [ 861.895487] env[62066]: _type = "Task" [ 861.895487] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.908699] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.938293] env[62066]: DEBUG nova.compute.utils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 861.944064] env[62066]: INFO nova.compute.claims [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.949320] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.949630] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.009395] env[62066]: DEBUG nova.policy [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53366fea3a6434fadadede81df16089', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33b756ddd30f4cb0b917fad171213266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.024211] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156280, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.208017] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156279, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.886986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.208565] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] da8b3bca-dbf5-4266-826d-ed93772fa04c/da8b3bca-dbf5-4266-826d-ed93772fa04c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.209869] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.209869] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64ab79f7-6d15-4b13-a616-b05c39d79829 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.217062] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 862.217062] env[62066]: value = "task-1156283" [ 862.217062] env[62066]: _type = "Task" [ 862.217062] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.228469] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156283, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.286580] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Successfully created port: dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.303468] env[62066]: DEBUG oslo_concurrency.lockutils [None req-620ba49a-8f55-4513-8d4f-72629f734417 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 35.497s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.311644] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156281, 'name': Rename_Task, 'duration_secs': 0.244338} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.311644] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.311644] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36880e93-02ac-4fc2-926f-0ef72f29efe8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.321351] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 862.321351] env[62066]: value = "task-1156284" [ 862.321351] env[62066]: _type = "Task" [ 862.321351] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.332710] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.389215] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.407291] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156282, 'name': ReconfigVM_Task, 'duration_secs': 0.305222} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.407591] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.407849] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f990f4a-29f0-4373-9e95-6632099ec04d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.415895] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 862.415895] env[62066]: value = "task-1156285" [ 862.415895] env[62066]: _type = "Task" [ 862.415895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.427795] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.449718] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.454065] env[62066]: INFO nova.compute.resource_tracker [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating resource usage from migration b2b4faf0-b29d-43d6-be2a-aeeeca362aa5 [ 862.524887] env[62066]: DEBUG oslo_vmware.api [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156280, 'name': PowerOnVM_Task, 'duration_secs': 0.835536} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.525327] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.525645] env[62066]: INFO nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Took 11.45 seconds to spawn the instance on the hypervisor. [ 862.525898] env[62066]: DEBUG nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.529691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d3e681-252c-4732-ad95-b2c97aa2352b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.657926] env[62066]: DEBUG nova.network.neutron [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Updating instance_info_cache with network_info: [{"id": "13c8bbfc-3aad-4881-966e-45eca3162138", "address": "fa:16:3e:35:52:68", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c8bbfc-3a", "ovs_interfaceid": "13c8bbfc-3aad-4881-966e-45eca3162138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.733493] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156283, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.733863] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.734761] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db292aa-82b7-4f41-bc1b-ecb6549bd977 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.769385] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] da8b3bca-dbf5-4266-826d-ed93772fa04c/da8b3bca-dbf5-4266-826d-ed93772fa04c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.774697] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb669303-ddba-4942-a75a-561997eca6ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.798085] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 862.798085] env[62066]: value = "task-1156286" [ 862.798085] env[62066]: _type = "Task" [ 862.798085] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.812515] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.833046] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.907691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2521858-6223-430c-8860-c23ff2940fba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.916607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d0c26e-f198-41c0-83bf-75d25e26a462 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.928775] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156285, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.954892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fba29b-897c-435d-bc6c-eb0ebe39a455 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.966496] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7141d6c0-de1d-4804-8f14-267989e2cb38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.981638] env[62066]: DEBUG nova.compute.provider_tree [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.053342] env[62066]: INFO nova.compute.manager [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Took 33.54 seconds to build instance. [ 863.160332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.160718] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Instance network_info: |[{"id": "13c8bbfc-3aad-4881-966e-45eca3162138", "address": "fa:16:3e:35:52:68", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c8bbfc-3a", "ovs_interfaceid": "13c8bbfc-3aad-4881-966e-45eca3162138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.161265] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:52:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '13c8bbfc-3aad-4881-966e-45eca3162138', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.170421] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating folder: Project (0c1a91ea6e0b4b2da6a16f327bc77a26). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.170734] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c19ebe3e-f5c3-4b7a-9fa7-8e94a43c8e4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.184888] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created folder: Project (0c1a91ea6e0b4b2da6a16f327bc77a26) in parent group-v251573. [ 863.185118] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating folder: Instances. Parent ref: group-v251716. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.185378] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eea71c86-e8fa-4c66-981c-53728de0da5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.196843] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created folder: Instances in parent group-v251716. [ 863.197130] env[62066]: DEBUG oslo.service.loopingcall [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.197359] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.197588] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15f17d4f-3090-47dc-8301-94c8f4b51b54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.219303] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.219303] env[62066]: value = "task-1156289" [ 863.219303] env[62066]: _type = "Task" [ 863.219303] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.227714] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156289, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.310973] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156286, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.330473] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156284, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.430991] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156285, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.465044] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.471365] env[62066]: DEBUG nova.compute.manager [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Received event network-changed-13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.471737] env[62066]: DEBUG nova.compute.manager [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Refreshing instance network info cache due to event network-changed-13c8bbfc-3aad-4881-966e-45eca3162138. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.472093] env[62066]: DEBUG oslo_concurrency.lockutils [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] Acquiring lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.472390] env[62066]: DEBUG oslo_concurrency.lockutils [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] Acquired lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.472619] env[62066]: DEBUG nova.network.neutron [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Refreshing network info cache for port 13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.484757] env[62066]: DEBUG nova.scheduler.client.report [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.502515] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='eab87e989d438f2bf2411d7fb4def551',container_format='bare',created_at=2024-10-07T00:57:35Z,direct_url=,disk_format='vmdk',id=b5438b44-e139-4d11-80c8-d6bdf1282d96,min_disk=1,min_ram=0,name='tempest-test-snap-947707190',owner='33b756ddd30f4cb0b917fad171213266',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-07T00:57:50Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.502822] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.502987] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.503772] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.503772] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.503772] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.503772] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.503931] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.504211] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.504280] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.504417] env[62066]: DEBUG nova.virt.hardware [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.505783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0caffa8-0aad-4b4d-b2f0-a74bafa6a19b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.516814] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c523fa2-6846-4880-ab80-f1a33803988a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.555704] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a19170dc-d2e1-4500-aede-c3eaf837ffb1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.053s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.730864] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156289, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.821792] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156286, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.834706] env[62066]: DEBUG oslo_vmware.api [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156284, 'name': PowerOnVM_Task, 'duration_secs': 1.218675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.835260] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.835628] env[62066]: INFO nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Took 10.23 seconds to spawn the instance on the hypervisor. [ 863.835992] env[62066]: DEBUG nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.837019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1a952b-8e1a-4801-b31b-33a8fc5632f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.862553] env[62066]: DEBUG nova.compute.manager [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Received event network-vif-plugged-dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.862553] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] Acquiring lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.862553] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.862553] env[62066]: DEBUG oslo_concurrency.lockutils [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.862553] env[62066]: DEBUG nova.compute.manager [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] No waiting events found dispatching network-vif-plugged-dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.862553] env[62066]: WARNING nova.compute.manager [req-f2426a5d-7402-4e08-8b1a-a9b6221c9b60 req-ef60cd85-01f2-4086-8709-79444969f67c service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Received unexpected event network-vif-plugged-dc192027-6e82-44a5-8f5c-95cbc8ea621c for instance with vm_state building and task_state spawning. [ 863.935098] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156285, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.994026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.558s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.994026] env[62066]: INFO nova.compute.manager [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Migrating [ 864.000447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.809s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.000904] env[62066]: DEBUG nova.objects.instance [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lazy-loading 'resources' on Instance uuid f3ef43e8-6092-44ac-8990-979810a9748f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.234103] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156289, 'name': CreateVM_Task, 'duration_secs': 0.639863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.234461] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.235281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.235551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.235986] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.236365] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72374fa0-fc35-473d-8c4e-8a6b069fb634 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.242372] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 864.242372] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52777505-cae5-8cf4-f04f-af31f2df5766" [ 864.242372] env[62066]: _type = "Task" [ 864.242372] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.253403] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52777505-cae5-8cf4-f04f-af31f2df5766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.308408] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Successfully updated port: dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.322256] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156286, 'name': ReconfigVM_Task, 'duration_secs': 1.140487} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.323880] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Reconfigured VM instance instance-0000004f to attach disk [datastore2] da8b3bca-dbf5-4266-826d-ed93772fa04c/da8b3bca-dbf5-4266-826d-ed93772fa04c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.325151] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74b24a08-38ce-40f6-a7bb-2d323d6b6cac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.333238] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 864.333238] env[62066]: value = "task-1156290" [ 864.333238] env[62066]: _type = "Task" [ 864.333238] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.343435] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156290, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.363156] env[62066]: INFO nova.compute.manager [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Took 33.96 seconds to build instance. [ 864.432352] env[62066]: DEBUG oslo_vmware.api [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156285, 'name': PowerOnVM_Task, 'duration_secs': 1.554013} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.432675] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.436423] env[62066]: DEBUG nova.compute.manager [None req-9e8b37f6-3538-439c-a6f5-500ac9dfb0f0 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 864.436618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c310a0a5-9788-4cc2-99b1-eaec4b648cd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.513244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.513476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.513695] env[62066]: DEBUG nova.network.neutron [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.756745] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52777505-cae5-8cf4-f04f-af31f2df5766, 'name': SearchDatastore_Task, 'duration_secs': 0.011605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.760190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.760190] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.760190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.760190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.760190] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.760190] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bdb7f3bf-d8ce-4d55-b974-49e00e0b68a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.771165] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.771367] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.772467] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c17eddd-b291-4e61-874a-cfcff77e4062 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.779192] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 864.779192] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a4afd2-73ee-0464-07b7-00bb3438f204" [ 864.779192] env[62066]: _type = "Task" [ 864.779192] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.795485] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a4afd2-73ee-0464-07b7-00bb3438f204, 'name': SearchDatastore_Task, 'duration_secs': 0.011412} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.795485] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0099675-b77b-4aa9-9c33-6f252c2e58f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.803301] env[62066]: DEBUG nova.network.neutron [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Updated VIF entry in instance network info cache for port 13c8bbfc-3aad-4881-966e-45eca3162138. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.803642] env[62066]: DEBUG nova.network.neutron [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Updating instance_info_cache with network_info: [{"id": "13c8bbfc-3aad-4881-966e-45eca3162138", "address": "fa:16:3e:35:52:68", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap13c8bbfc-3a", "ovs_interfaceid": "13c8bbfc-3aad-4881-966e-45eca3162138", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.806435] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 864.806435] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5261651a-b0bf-df0a-b474-af1336e4061e" [ 864.806435] env[62066]: _type = "Task" [ 864.806435] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.817747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.817849] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.819335] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.819335] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5261651a-b0bf-df0a-b474-af1336e4061e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.846667] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156290, 'name': Rename_Task, 'duration_secs': 0.375212} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.846995] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.847392] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-064e1319-67b6-4070-a5b8-db69e70d4cdb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.858312] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 864.858312] env[62066]: value = "task-1156291" [ 864.858312] env[62066]: _type = "Task" [ 864.858312] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.868791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bc47f821-3719-4ab4-b77c-25648496d7bf tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.485s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.869427] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156291, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.871157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe3b6e7-f0a3-47e4-92cb-db8d27359088 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.879095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4455d2a-7f0a-4939-9ae3-a8dd0b53f140 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.914090] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e8b4d3-3681-4dae-b0a8-cc8711b927c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.923891] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7f043b-e35f-4f99-92bc-78c67bfa2d63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.940611] env[62066]: DEBUG nova.compute.provider_tree [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.203410] env[62066]: DEBUG nova.compute.manager [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-changed-0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.203637] env[62066]: DEBUG nova.compute.manager [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing instance network info cache due to event network-changed-0d8007bd-9e20-4780-a21e-a22c8c7dac13. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.203885] env[62066]: DEBUG oslo_concurrency.lockutils [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.204113] env[62066]: DEBUG oslo_concurrency.lockutils [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.204342] env[62066]: DEBUG nova.network.neutron [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing network info cache for port 0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.307666] env[62066]: DEBUG oslo_concurrency.lockutils [req-f808f824-f003-4844-8d31-e483730ff946 req-222f0add-b908-4770-b655-79a24c8a8e11 service nova] Releasing lock "refresh_cache-3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.318601] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5261651a-b0bf-df0a-b474-af1336e4061e, 'name': SearchDatastore_Task, 'duration_secs': 0.010896} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.318915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.319224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d/3b712d48-1365-4eed-8d6d-8ee6dbf51c2d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.319547] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b937386-e029-4a4d-bbd6-50ed32e6d247 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.332219] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 865.332219] env[62066]: value = "task-1156292" [ 865.332219] env[62066]: _type = "Task" [ 865.332219] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.341815] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.354548] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.374847] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156291, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.443963] env[62066]: DEBUG nova.scheduler.client.report [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.527413] env[62066]: DEBUG nova.network.neutron [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.531534] env[62066]: DEBUG nova.network.neutron [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Updating instance_info_cache with network_info: [{"id": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "address": "fa:16:3e:59:bc:5a", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc192027-6e", "ovs_interfaceid": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.735318] env[62066]: DEBUG nova.compute.manager [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Received event network-changed-dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.735318] env[62066]: DEBUG nova.compute.manager [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Refreshing instance network info cache due to event network-changed-dc192027-6e82-44a5-8f5c-95cbc8ea621c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.735318] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] Acquiring lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.846569] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156292, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494516} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.847118] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d/3b712d48-1365-4eed-8d6d-8ee6dbf51c2d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.847606] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.848057] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20f74eeb-2321-4a17-9ff5-9eb676c4886c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.857241] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 865.857241] env[62066]: value = "task-1156293" [ 865.857241] env[62066]: _type = "Task" [ 865.857241] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.878940] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.886628] env[62066]: DEBUG oslo_vmware.api [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156291, 'name': PowerOnVM_Task, 'duration_secs': 0.599988} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.887457] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.887850] env[62066]: INFO nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Took 9.65 seconds to spawn the instance on the hypervisor. [ 865.888232] env[62066]: DEBUG nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.890157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4e6a09-b7d9-495d-9298-f11d0143561a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.952825] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.955988] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 23.123s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.969274] env[62066]: DEBUG nova.network.neutron [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updated VIF entry in instance network info cache for port 0d8007bd-9e20-4780-a21e-a22c8c7dac13. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.969715] env[62066]: DEBUG nova.network.neutron [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.977154] env[62066]: INFO nova.scheduler.client.report [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleted allocations for instance f3ef43e8-6092-44ac-8990-979810a9748f [ 866.029489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.043018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.043018] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Instance network_info: |[{"id": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "address": "fa:16:3e:59:bc:5a", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc192027-6e", "ovs_interfaceid": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.043018] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] Acquired lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.043018] env[62066]: DEBUG nova.network.neutron [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Refreshing network info cache for port dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.043018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:bc:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc192027-6e82-44a5-8f5c-95cbc8ea621c', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.051883] env[62066]: DEBUG oslo.service.loopingcall [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.053654] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.054476] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-742bd550-55dc-4490-ae4f-2f1ac3f9d519 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.078738] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.078738] env[62066]: value = "task-1156294" [ 866.078738] env[62066]: _type = "Task" [ 866.078738] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.088739] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156294, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.369706] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.373023] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.373023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2299cc-0233-40e2-bf57-47169c7ba4f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.379207] env[62066]: INFO nova.compute.manager [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Unrescuing [ 866.379207] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.379207] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.379207] env[62066]: DEBUG nova.network.neutron [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.414379] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d/3b712d48-1365-4eed-8d6d-8ee6dbf51c2d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.420371] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68bedb22-8eaa-40bc-932f-be64e938ca66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.439904] env[62066]: INFO nova.compute.manager [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Took 36.03 seconds to build instance. [ 866.445710] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 866.445710] env[62066]: value = "task-1156295" [ 866.445710] env[62066]: _type = "Task" [ 866.445710] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.457901] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.462871] env[62066]: INFO nova.compute.claims [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.473570] env[62066]: DEBUG oslo_concurrency.lockutils [req-d86b1d2f-3951-42db-b310-bd3aaac6b63f req-f00290bf-54e4-4385-a83f-ffdb4046ef18 service nova] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.486377] env[62066]: DEBUG oslo_concurrency.lockutils [None req-902be719-8767-4697-9931-1af9f44e244f tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "f3ef43e8-6092-44ac-8990-979810a9748f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.293s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.595107] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156294, 'name': CreateVM_Task, 'duration_secs': 0.35954} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.595296] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.596128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.596308] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.596696] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.596952] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56a710e2-3520-453f-aec9-ef6fb8d7b162 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.602239] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 866.602239] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cb99de-2118-a2e3-cbe7-cb5ab1c710a9" [ 866.602239] env[62066]: _type = "Task" [ 866.602239] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.611457] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cb99de-2118-a2e3-cbe7-cb5ab1c710a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.787669] env[62066]: DEBUG nova.network.neutron [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Updated VIF entry in instance network info cache for port dc192027-6e82-44a5-8f5c-95cbc8ea621c. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 866.788122] env[62066]: DEBUG nova.network.neutron [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Updating instance_info_cache with network_info: [{"id": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "address": "fa:16:3e:59:bc:5a", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc192027-6e", "ovs_interfaceid": "dc192027-6e82-44a5-8f5c-95cbc8ea621c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.947453] env[62066]: DEBUG oslo_concurrency.lockutils [None req-78d22fc8-8649-4253-b350-989851b528b4 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.562s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.960126] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156295, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.969248] env[62066]: INFO nova.compute.resource_tracker [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating resource usage from migration 45cece1f-bf1a-4a05-81c8-de8c7a858bdd [ 867.038297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.038572] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.038777] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.038965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.039171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.047056] env[62066]: INFO nova.compute.manager [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Terminating instance [ 867.048928] env[62066]: DEBUG nova.compute.manager [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.049149] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 867.050676] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8c9ac4-85e8-429e-9e3c-6f51f431f601 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.065073] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 867.067599] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cac0d96-5e1d-4227-9ec4-c81a43251a13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.076083] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 867.076083] env[62066]: value = "task-1156296" [ 867.076083] env[62066]: _type = "Task" [ 867.076083] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.088848] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.114136] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.117028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Processing image b5438b44-e139-4d11-80c8-d6bdf1282d96 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.117028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.117028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.117028] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.117028] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a7a1ab5-1f76-4b72-a845-69c2f8e1d042 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.126121] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.126352] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.127265] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e912b327-5ac2-4694-95d8-bfabc8a11a1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.134652] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 867.134652] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f07951-7347-1908-24a7-3689bb3e5c4e" [ 867.134652] env[62066]: _type = "Task" [ 867.134652] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.144625] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f07951-7347-1908-24a7-3689bb3e5c4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.145712] env[62066]: DEBUG nova.network.neutron [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.291075] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef45c090-d835-4c93-a2ef-44232645d848 req-c12be5d1-81ca-46e2-91c4-27af39e3afe4 service nova] Releasing lock "refresh_cache-4914be5d-2cc8-48b7-96e5-9192e5c73fa3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.324305] env[62066]: DEBUG nova.compute.manager [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Received event network-changed-6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.324696] env[62066]: DEBUG nova.compute.manager [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Refreshing instance network info cache due to event network-changed-6244eb3b-c073-494b-b711-f8712351da75. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.325044] env[62066]: DEBUG oslo_concurrency.lockutils [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] Acquiring lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.325318] env[62066]: DEBUG oslo_concurrency.lockutils [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] Acquired lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.325588] env[62066]: DEBUG nova.network.neutron [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Refreshing network info cache for port 6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.373487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b38dea-b457-43d5-9ab5-4d943cbf1833 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.381763] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b3faa7-5d96-4554-83e9-5abaf73ebcd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.413043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903c478b-78d1-4c1d-ad32-d500d252634b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.420989] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14b66ff-5322-4fba-bd38-7cd1ee718c18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.437074] env[62066]: DEBUG nova.compute.provider_tree [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.458697] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156295, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.464689] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.464945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.465166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.465352] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.465521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.468025] env[62066]: INFO nova.compute.manager [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Terminating instance [ 867.469929] env[62066]: DEBUG nova.compute.manager [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.469929] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 867.470575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483f3e0c-7775-425e-9266-d7d284c54267 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.478972] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 867.479242] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b915c3e-dc87-4621-a48f-5798358b8804 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.487125] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 867.487125] env[62066]: value = "task-1156297" [ 867.487125] env[62066]: _type = "Task" [ 867.487125] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.495512] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156297, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.549194] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e444452b-f597-409d-b58a-91570a9edb86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.569473] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 867.586466] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156296, 'name': PowerOffVM_Task, 'duration_secs': 0.254252} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.586704] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.586879] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.587152] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8a11390-d201-4861-ae07-3cb9d592cef7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.645716] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 867.646058] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Fetch image to [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531/OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 867.646275] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Downloading stream optimized image b5438b44-e139-4d11-80c8-d6bdf1282d96 to [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531/OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531.vmdk on the data store datastore1 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 867.646456] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Downloading image file data b5438b44-e139-4d11-80c8-d6bdf1282d96 to the ESX as VM named 'OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 867.648783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.649437] env[62066]: DEBUG nova.objects.instance [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'flavor' on Instance uuid 831c0bed-8a41-4672-aa08-d60b64f365b8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.652829] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.653191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.653404] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleting the datastore file [datastore2] de9eb96d-e71b-4878-83b8-f75966dc3f48 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.653914] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6e567f5-ecb6-4be0-bae6-a47468b1190b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.661016] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for the task: (returnval){ [ 867.661016] env[62066]: value = "task-1156299" [ 867.661016] env[62066]: _type = "Task" [ 867.661016] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.673671] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156299, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.727505] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 867.727505] env[62066]: value = "resgroup-9" [ 867.727505] env[62066]: _type = "ResourcePool" [ 867.727505] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 867.727831] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-afa5ab9e-831a-4fcb-aab3-3ba6f0621e9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.760637] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease: (returnval){ [ 867.760637] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 867.760637] env[62066]: _type = "HttpNfcLease" [ 867.760637] env[62066]: } obtained for vApp import into resource pool (val){ [ 867.760637] env[62066]: value = "resgroup-9" [ 867.760637] env[62066]: _type = "ResourcePool" [ 867.760637] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 867.760999] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the lease: (returnval){ [ 867.760999] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 867.760999] env[62066]: _type = "HttpNfcLease" [ 867.760999] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 867.769518] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 867.769518] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 867.769518] env[62066]: _type = "HttpNfcLease" [ 867.769518] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 867.940207] env[62066]: DEBUG nova.scheduler.client.report [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.959707] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156295, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.000321] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156297, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.064707] env[62066]: DEBUG nova.network.neutron [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updated VIF entry in instance network info cache for port 6244eb3b-c073-494b-b711-f8712351da75. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.065124] env[62066]: DEBUG nova.network.neutron [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.076819] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.077185] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72b45c65-7eba-48ec-8013-0f3fb618a890 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.085762] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 868.085762] env[62066]: value = "task-1156301" [ 868.085762] env[62066]: _type = "Task" [ 868.085762] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.097844] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 868.098055] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 868.158212] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dcc7b7-751f-4674-a3dc-09b60a9275d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.173094] env[62066]: DEBUG oslo_vmware.api [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Task: {'id': task-1156299, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182794} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.192781] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.193017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 868.193213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 868.193392] env[62066]: INFO nova.compute.manager [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Took 1.14 seconds to destroy the instance on the hypervisor. [ 868.193643] env[62066]: DEBUG oslo.service.loopingcall [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.193970] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.194285] env[62066]: DEBUG nova.compute.manager [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.194385] env[62066]: DEBUG nova.network.neutron [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.196113] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dabe821a-3184-46f6-87f4-7e78bf2c7231 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.205084] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 868.205084] env[62066]: value = "task-1156302" [ 868.205084] env[62066]: _type = "Task" [ 868.205084] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.221154] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.271327] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.271327] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 868.271327] env[62066]: _type = "HttpNfcLease" [ 868.271327] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 868.451026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.494s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.451026] env[62066]: INFO nova.compute.manager [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Migrating [ 868.458208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.585s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.460473] env[62066]: INFO nova.compute.claims [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.488956] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156295, 'name': ReconfigVM_Task, 'duration_secs': 1.886736} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.495186] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d/3b712d48-1365-4eed-8d6d-8ee6dbf51c2d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.495186] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51fedfe8-2858-4277-84d7-638725d3e670 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.505505] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156297, 'name': PowerOffVM_Task, 'duration_secs': 0.991392} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.510039] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.510039] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.510039] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 868.510039] env[62066]: value = "task-1156303" [ 868.510039] env[62066]: _type = "Task" [ 868.510039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.510239] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3241deaa-8f39-4d7b-a5e8-93ddfbcff85a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.524793] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156303, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.568496] env[62066]: DEBUG oslo_concurrency.lockutils [req-86b5db0c-a347-4387-8b7b-703dd95298dd req-166e08c9-ab30-4247-88ac-34580e09bcc6 service nova] Releasing lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.579264] env[62066]: DEBUG nova.compute.manager [req-742dd1ba-f1db-40ae-a1b1-9e30289ea536 req-5489b34a-90d9-4bee-aa17-028d92c67daa service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Received event network-vif-deleted-7d4d5170-fd58-4398-b403-887f40e3c29c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.579627] env[62066]: INFO nova.compute.manager [req-742dd1ba-f1db-40ae-a1b1-9e30289ea536 req-5489b34a-90d9-4bee-aa17-028d92c67daa service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Neutron deleted interface 7d4d5170-fd58-4398-b403-887f40e3c29c; detaching it from the instance and deleting it from the info cache [ 868.579849] env[62066]: DEBUG nova.network.neutron [req-742dd1ba-f1db-40ae-a1b1-9e30289ea536 req-5489b34a-90d9-4bee-aa17-028d92c67daa service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.586351] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 868.586687] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 868.586980] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Deleting the datastore file [datastore2] da8b3bca-dbf5-4266-826d-ed93772fa04c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.587587] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f8b2871-bdf8-495c-a4b1-fa9cb8fd050f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.595511] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for the task: (returnval){ [ 868.595511] env[62066]: value = "task-1156305" [ 868.595511] env[62066]: _type = "Task" [ 868.595511] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.605251] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.607051] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.607414] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.612816] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e77fbf1-05d1-4256-bee3-a21b378ccbeb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.629360] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156305, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.630843] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 868.630843] env[62066]: value = "task-1156306" [ 868.630843] env[62066]: _type = "Task" [ 868.630843] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.639674] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156306, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.715827] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156302, 'name': PowerOffVM_Task, 'duration_secs': 0.273409} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.716207] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.721579] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfiguring VM instance instance-00000043 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 868.725044] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d30eaadb-b443-4350-932a-ff30b6f89ea5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.745244] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 868.745244] env[62066]: value = "task-1156307" [ 868.745244] env[62066]: _type = "Task" [ 868.745244] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.754656] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156307, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.772023] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.772023] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 868.772023] env[62066]: _type = "HttpNfcLease" [ 868.772023] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 868.772023] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 868.772023] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526fd723-7a45-43b9-7e48-a975a4ef906a" [ 868.772023] env[62066]: _type = "HttpNfcLease" [ 868.772023] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 868.772023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dece24-62aa-4f8b-bd91-5e8f2eba41da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.782378] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 868.783063] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 868.847602] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a01a4516-5ed6-4451-b77a-77a7d0a60907 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.980115] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.981069] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.981069] env[62066]: DEBUG nova.network.neutron [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.010481] env[62066]: DEBUG nova.network.neutron [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.022519] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156303, 'name': Rename_Task, 'duration_secs': 0.180182} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.023380] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.023652] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-623ec4ee-ef6a-4d98-bfa5-1834dc2085dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.032015] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 869.032015] env[62066]: value = "task-1156308" [ 869.032015] env[62066]: _type = "Task" [ 869.032015] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.040620] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.083505] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6f66aea-afef-489c-b25e-4b8739edb092 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.093866] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa750fb-1759-467c-bf73-d67b0e22f420 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.113401] env[62066]: DEBUG oslo_vmware.api [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Task: {'id': task-1156305, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.113654] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.113864] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.114077] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.114264] env[62066]: INFO nova.compute.manager [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Took 1.64 seconds to destroy the instance on the hypervisor. [ 869.114507] env[62066]: DEBUG oslo.service.loopingcall [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.114698] env[62066]: DEBUG nova.compute.manager [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.114790] env[62066]: DEBUG nova.network.neutron [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.133512] env[62066]: DEBUG nova.compute.manager [req-742dd1ba-f1db-40ae-a1b1-9e30289ea536 req-5489b34a-90d9-4bee-aa17-028d92c67daa service nova] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Detach interface failed, port_id=7d4d5170-fd58-4398-b403-887f40e3c29c, reason: Instance de9eb96d-e71b-4878-83b8-f75966dc3f48 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 869.142647] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156306, 'name': ReconfigVM_Task, 'duration_secs': 0.149185} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.142957] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 869.259778] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156307, 'name': ReconfigVM_Task, 'duration_secs': 0.497999} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.263267] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfigured VM instance instance-00000043 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 869.263430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.263723] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd4deb42-b585-4d0d-9cea-116297ac6994 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.272940] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 869.272940] env[62066]: value = "task-1156309" [ 869.272940] env[62066]: _type = "Task" [ 869.272940] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.285088] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.516236] env[62066]: INFO nova.compute.manager [-] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Took 1.32 seconds to deallocate network for instance. [ 869.544073] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156308, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.649849] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.650183] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.650280] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.650467] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.650617] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.650788] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.651191] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.651407] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.651636] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.651819] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.651999] env[62066]: DEBUG nova.virt.hardware [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.658154] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfiguring VM instance instance-00000030 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 869.659171] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f291387-ba13-4ea7-9bad-a27823bfbef6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.692546] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 869.692546] env[62066]: value = "task-1156310" [ 869.692546] env[62066]: _type = "Task" [ 869.692546] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.705329] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.731488] env[62066]: DEBUG nova.network.neutron [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.788563] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156309, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.854673] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 869.854954] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 869.856434] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a1c4a3-28a7-4ce5-bb5c-8d5f40c4ac27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.865285] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 869.865547] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 869.865894] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1b19d5fc-cb1f-4b72-926e-34c0fa14de31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.923594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00aaa66-ebd6-4516-a7d7-ffc6fdd7ac5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.927107] env[62066]: DEBUG nova.network.neutron [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.933570] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e09220-311b-4e42-b7be-bd63b4dacb5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.968599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fb7bb2-b624-409f-ab2e-b509e6613b64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.978232] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780e5d2e-2932-433f-a466-b173701ab334 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.994594] env[62066]: DEBUG nova.compute.provider_tree [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.023337] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.042898] env[62066]: DEBUG oslo_vmware.api [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156308, 'name': PowerOnVM_Task, 'duration_secs': 0.554385} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.043044] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.044026] env[62066]: INFO nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Took 9.15 seconds to spawn the instance on the hypervisor. [ 870.044026] env[62066]: DEBUG nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.044243] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4963bf-eb8b-4a00-81d7-124d405e4d40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.203455] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156310, 'name': ReconfigVM_Task, 'duration_secs': 0.223739} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.203595] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfigured VM instance instance-00000030 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 870.204368] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f00799-9e80-405d-b27c-fba54450ed1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.228086] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.228418] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-250977f2-48d8-44ec-bd41-74bf47361750 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.245246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.257248] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 870.257248] env[62066]: value = "task-1156311" [ 870.257248] env[62066]: _type = "Task" [ 870.257248] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.269788] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.286830] env[62066]: DEBUG oslo_vmware.api [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156309, 'name': PowerOnVM_Task, 'duration_secs': 0.594378} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.287331] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.287586] env[62066]: DEBUG nova.compute.manager [None req-89e61d38-ca0d-4f58-b21f-4460dfbbf426 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.288618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9d6ad9-13b5-4159-90e8-b480c9dcbad7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.318709] env[62066]: DEBUG oslo_vmware.rw_handles [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207103e-9fb9-e8e8-1416-f8199321b093/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 870.318709] env[62066]: INFO nova.virt.vmwareapi.images [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Downloaded image file data b5438b44-e139-4d11-80c8-d6bdf1282d96 [ 870.319697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5946c27b-ef3d-4fd0-a24d-0c71a7f22c58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.338801] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a23dfe7f-9655-47b7-a44b-f9aaea4ac0e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.366532] env[62066]: INFO nova.virt.vmwareapi.images [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] The imported VM was unregistered [ 870.368984] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 870.369254] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating directory with path [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.369558] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79c25a30-92ca-409f-a55d-e920f77fd80d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.381482] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created directory with path [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.381766] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531/OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531.vmdk to [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 870.382137] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-bfd8fe8a-8314-4d28-af20-462b0bd72504 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.390833] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 870.390833] env[62066]: value = "task-1156313" [ 870.390833] env[62066]: _type = "Task" [ 870.390833] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.400661] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.429999] env[62066]: INFO nova.compute.manager [-] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Took 1.31 seconds to deallocate network for instance. [ 870.498725] env[62066]: DEBUG nova.scheduler.client.report [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.563403] env[62066]: INFO nova.compute.manager [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Took 35.27 seconds to build instance. [ 870.769999] env[62066]: DEBUG oslo_vmware.api [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156311, 'name': ReconfigVM_Task, 'duration_secs': 0.477519} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.770440] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Reconfigured VM instance instance-00000030 to attach disk [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1/ccb9f50f-dcc3-4d81-944e-d70803185ae1.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.770851] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 870.796853] env[62066]: DEBUG nova.compute.manager [req-38f25cc7-94ce-47c9-a211-6d6924be3564 req-68ce2109-8323-4366-8c96-b722f012fd15 service nova] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Received event network-vif-deleted-ec0e826d-c94d-4295-a202-c1e2bfdcefd5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.902389] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.937423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.005486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.006166] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.010209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.120s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.011787] env[62066]: INFO nova.compute.claims [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.065125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-82761842-0f3c-4415-b820-75b2c9d593e1 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.789s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.279032] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a90d1c-c218-4f7b-be8c-a255cb1572ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.303966] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2a4041-8aa2-4c3d-ba75-b274dd38bfbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.324547] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.403762] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.512390] env[62066]: DEBUG nova.compute.utils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 871.517647] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.517647] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 871.535982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.536912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.536912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.537310] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.537544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.540933] env[62066]: INFO nova.compute.manager [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Terminating instance [ 871.543263] env[62066]: DEBUG nova.compute.manager [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.543476] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.545151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37be01bb-3537-443b-bf81-63c41fff9a61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.555136] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.555465] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0653427b-4053-413f-b2b2-fcf8ca980f69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.568285] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 871.568285] env[62066]: value = "task-1156314" [ 871.568285] env[62066]: _type = "Task" [ 871.568285] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.578796] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.583771] env[62066]: DEBUG nova.policy [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1984a406905949b1b7653c4b235ccc23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b83b0c5ef67b4135b71299dfe2ec4263', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.763359] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb46710c-b118-4640-bec7-fd9c7ec68695 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.789203] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.906036] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.907969] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Successfully created port: fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.964016] env[62066]: DEBUG nova.network.neutron [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Port 81a78519-507f-4de8-8e08-3a469f901c85 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 872.018443] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.085568] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156314, 'name': PowerOffVM_Task, 'duration_secs': 0.261063} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.089917] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.090559] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.091922] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7ae30fc-a6c5-4cbf-8d09-26d5431aebc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.168329] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.168805] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.169060] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore1] 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.169552] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24cfd0e3-2473-4edf-a40e-e04e588dcc12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.181450] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 872.181450] env[62066]: value = "task-1156316" [ 872.181450] env[62066]: _type = "Task" [ 872.181450] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.199116] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.253239] env[62066]: DEBUG nova.compute.manager [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.253813] env[62066]: DEBUG nova.compute.manager [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing instance network info cache due to event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.254038] env[62066]: DEBUG oslo_concurrency.lockutils [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.254810] env[62066]: DEBUG oslo_concurrency.lockutils [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.255166] env[62066]: DEBUG nova.network.neutron [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.296583] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.297568] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6853b262-af5c-4fc9-95d4-c988a07e50d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.311940] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 872.311940] env[62066]: value = "task-1156317" [ 872.311940] env[62066]: _type = "Task" [ 872.311940] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.325077] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.407415] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.532548] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f73c945-ba4c-4e0e-837c-83aa8e2e5a13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.543082] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a4e2aa-bb5d-4b04-8212-c4cd20956557 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.578558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed3d702-20b5-4710-8ebe-628cf277f291 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.588310] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43493918-22f6-4284-9f61-8eb0f8c93d27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.606929] env[62066]: DEBUG nova.compute.provider_tree [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.695317] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.831699] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.904511] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156313, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.511293} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.904812] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531/OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531.vmdk to [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk. [ 872.905018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Cleaning up location [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 872.905193] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_b69728f4-9233-479f-80c8-8011b530a531 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.905473] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-704b6af8-c265-437e-913a-1aca4d1f03c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.913897] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 872.913897] env[62066]: value = "task-1156318" [ 872.913897] env[62066]: _type = "Task" [ 872.913897] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.925825] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.994297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.994679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.994964] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.038223] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.065492] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.065861] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.066068] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.066268] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.066513] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.066715] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.066961] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.067203] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.067465] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.067694] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.067924] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.070316] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ee11a8-0c9a-4d2f-bffe-8e3d5a170664 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.080412] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da932f5-a012-45c6-b1d2-7718da4a93c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.109906] env[62066]: DEBUG nova.scheduler.client.report [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.180957] env[62066]: DEBUG nova.network.neutron [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updated VIF entry in instance network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.181402] env[62066]: DEBUG nova.network.neutron [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.196062] env[62066]: DEBUG oslo_vmware.api [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.818433} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.196458] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.196662] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.196841] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.197027] env[62066]: INFO nova.compute.manager [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Took 1.65 seconds to destroy the instance on the hypervisor. [ 873.197312] env[62066]: DEBUG oslo.service.loopingcall [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.197751] env[62066]: DEBUG nova.compute.manager [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.197847] env[62066]: DEBUG nova.network.neutron [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 873.324456] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156317, 'name': PowerOffVM_Task, 'duration_secs': 0.667439} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.324689] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.324959] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 873.428215] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.428501] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.428674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.428924] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk to [datastore1] 4914be5d-2cc8-48b7-96e5-9192e5c73fa3/4914be5d-2cc8-48b7-96e5-9192e5c73fa3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.429975] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-400433ff-d3e9-46d7-9b4b-88034e57cfac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.437882] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 873.437882] env[62066]: value = "task-1156319" [ 873.437882] env[62066]: _type = "Task" [ 873.437882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.446638] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.619041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.620939] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.624304] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.695s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.624901] env[62066]: INFO nova.compute.claims [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.634090] env[62066]: DEBUG nova.compute.manager [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Received event network-vif-plugged-fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.634309] env[62066]: DEBUG oslo_concurrency.lockutils [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] Acquiring lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.634519] env[62066]: DEBUG oslo_concurrency.lockutils [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.634686] env[62066]: DEBUG oslo_concurrency.lockutils [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.635267] env[62066]: DEBUG nova.compute.manager [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] No waiting events found dispatching network-vif-plugged-fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.635267] env[62066]: WARNING nova.compute.manager [req-600b2a9a-d3c8-450c-83b3-05d924f3b3f4 req-95001678-a4d2-46ea-a945-1418f8daef81 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Received unexpected event network-vif-plugged-fb8eff91-5d00-4cdc-87e0-8aa48d429972 for instance with vm_state building and task_state spawning. [ 873.684616] env[62066]: DEBUG oslo_concurrency.lockutils [req-61c94abe-0ff7-4624-bc2c-45b5437c3fbc req-2c9cce57-59f0-4698-b8c8-bb08ba2ca42b service nova] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.728083] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Successfully updated port: fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.831398] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.831660] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.831844] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.832031] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.832185] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.832335] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.832541] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.832705] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.832875] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.833046] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.833230] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.838142] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d875a73-10f0-451c-8239-a90e2945a755 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.854351] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 873.854351] env[62066]: value = "task-1156320" [ 873.854351] env[62066]: _type = "Task" [ 873.854351] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.863607] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.950240] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.031737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.031959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.032172] env[62066]: DEBUG nova.network.neutron [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.131365] env[62066]: DEBUG nova.compute.utils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 874.136221] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.136415] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.139047] env[62066]: DEBUG nova.network.neutron [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.185273] env[62066]: DEBUG nova.policy [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1984a406905949b1b7653c4b235ccc23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b83b0c5ef67b4135b71299dfe2ec4263', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.231674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.231826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.232072] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.294436] env[62066]: DEBUG nova.compute.manager [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.295027] env[62066]: DEBUG nova.compute.manager [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing instance network info cache due to event network-changed-4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.295027] env[62066]: DEBUG oslo_concurrency.lockutils [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] Acquiring lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.295027] env[62066]: DEBUG oslo_concurrency.lockutils [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] Acquired lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.295359] env[62066]: DEBUG nova.network.neutron [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Refreshing network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.366383] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156320, 'name': ReconfigVM_Task, 'duration_secs': 0.234767} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.366781] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 874.451849] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.480480] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Successfully created port: 55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.638022] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.642021] env[62066]: INFO nova.compute.manager [-] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Took 1.44 seconds to deallocate network for instance. [ 874.778339] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.875059] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.875405] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.876531] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.876806] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.877043] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.877217] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.877480] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.877652] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.877835] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.878075] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.878276] env[62066]: DEBUG nova.virt.hardware [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.887997] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfiguring VM instance instance-0000004a to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 874.893770] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52296cdc-5662-42fc-b591-1960ab964a77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.918421] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 874.918421] env[62066]: value = "task-1156321" [ 874.918421] env[62066]: _type = "Task" [ 874.918421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.931323] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156321, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.950465] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.954691] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Updating instance_info_cache with network_info: [{"id": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "address": "fa:16:3e:8f:59:4f", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb8eff91-5d", "ovs_interfaceid": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.156506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.157935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a170f7-10ed-4894-9c7f-908c579c4edd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.172388] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2406084-8713-45d4-9923-9398127df279 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.217531] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b4b8c4-b4ba-4b2a-ae79-e193e38bdc19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.233092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62632db-e780-4ea9-820e-de3803598a86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.253709] env[62066]: DEBUG nova.compute.provider_tree [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.429918] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156321, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.451379] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.457219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.457689] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Instance network_info: |[{"id": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "address": "fa:16:3e:8f:59:4f", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb8eff91-5d", "ovs_interfaceid": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.458201] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:59:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb8eff91-5d00-4cdc-87e0-8aa48d429972', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.470034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Creating folder: Project (b83b0c5ef67b4135b71299dfe2ec4263). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.470034] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4202119-085f-4592-903e-65bf038953d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.481025] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Created folder: Project (b83b0c5ef67b4135b71299dfe2ec4263) in parent group-v251573. [ 875.485622] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Creating folder: Instances. Parent ref: group-v251721. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.486191] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc8c7a57-3883-48bc-a107-57f197400269 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.490959] env[62066]: DEBUG nova.network.neutron [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updated VIF entry in instance network info cache for port 4e3a7a96-3bfe-4fbe-9515-fd341f054700. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 875.491950] env[62066]: DEBUG nova.network.neutron [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [{"id": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "address": "fa:16:3e:27:6c:0e", "network": {"id": "334d0c18-636d-4664-857f-5d89cdfc650f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1363475790-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93c0e92cfec42f4b4a20e9fb4a32088", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271a82f1-1d09-4ad3-9c15-07269bad114c", "external-id": "nsx-vlan-transportzone-441", "segmentation_id": 441, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e3a7a96-3b", "ovs_interfaceid": "4e3a7a96-3bfe-4fbe-9515-fd341f054700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.497580] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Created folder: Instances in parent group-v251721. [ 875.498154] env[62066]: DEBUG oslo.service.loopingcall [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.498279] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.498539] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98bd9d67-967e-4c46-aece-9c5abba6ff52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.521805] env[62066]: DEBUG nova.network.neutron [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.528289] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.528289] env[62066]: value = "task-1156324" [ 875.528289] env[62066]: _type = "Task" [ 875.528289] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.542748] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156324, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.648724] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.679500] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.679818] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.679993] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.680254] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.680366] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.680519] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.680738] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.680908] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.681106] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.681283] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.681466] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.682422] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3237e882-0567-4f5c-8c83-b88eeb96beb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.691658] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4d7050-0399-49f1-9c9a-75f6fedb428a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.761864] env[62066]: DEBUG nova.scheduler.client.report [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.878596] env[62066]: DEBUG nova.compute.manager [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Received event network-changed-fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.878898] env[62066]: DEBUG nova.compute.manager [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Refreshing instance network info cache due to event network-changed-fb8eff91-5d00-4cdc-87e0-8aa48d429972. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.879233] env[62066]: DEBUG oslo_concurrency.lockutils [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] Acquiring lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.879455] env[62066]: DEBUG oslo_concurrency.lockutils [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] Acquired lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.879763] env[62066]: DEBUG nova.network.neutron [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Refreshing network info cache for port fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.928993] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156321, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.954517] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.995217] env[62066]: DEBUG oslo_concurrency.lockutils [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] Releasing lock "refresh_cache-831c0bed-8a41-4672-aa08-d60b64f365b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.995660] env[62066]: DEBUG nova.compute.manager [req-9f90076e-a065-4432-b23e-9d87feaf4efe req-6a5d3424-a95c-4608-a8d6-5e77ecd8691f service nova] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Received event network-vif-deleted-13c8bbfc-3aad-4881-966e-45eca3162138 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.022664] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.040313] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156324, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.110709] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Successfully updated port: 55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.268398] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.269105] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.271941] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.542s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.272188] env[62066]: DEBUG nova.objects.instance [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 876.431966] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156321, 'name': ReconfigVM_Task, 'duration_secs': 1.18802} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.432394] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfigured VM instance instance-0000004a to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 876.433336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19cdd4c-c092-4ee8-8099-7be8dcfa6110 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.461856] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.466359] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9463352e-b250-43a2-b51e-bbceb6d3cf63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.489976] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156319, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.916686} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.491420] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b5438b44-e139-4d11-80c8-d6bdf1282d96/b5438b44-e139-4d11-80c8-d6bdf1282d96.vmdk to [datastore1] 4914be5d-2cc8-48b7-96e5-9192e5c73fa3/4914be5d-2cc8-48b7-96e5-9192e5c73fa3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.491907] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 876.491907] env[62066]: value = "task-1156325" [ 876.491907] env[62066]: _type = "Task" [ 876.491907] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.492762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e05bfc0-a21f-423b-af65-4a0044644e7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.514423] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156325, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.529701] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 4914be5d-2cc8-48b7-96e5-9192e5c73fa3/4914be5d-2cc8-48b7-96e5-9192e5c73fa3.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.538059] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b11f7cc2-af8c-439b-a4a2-a9a2efdae124 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.583369] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156324, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.587102] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 876.587102] env[62066]: value = "task-1156326" [ 876.587102] env[62066]: _type = "Task" [ 876.587102] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.595509] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156326, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.600186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f238f2a-81fe-4a7e-8d69-082bb7bbaeae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.619011] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.619207] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.619352] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.621278] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99051306-915e-4e90-a96b-d77f78273995 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.629108] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 876.657364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.658209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.658209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.658394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.658702] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.664269] env[62066]: INFO nova.compute.manager [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Terminating instance [ 876.665425] env[62066]: DEBUG nova.compute.manager [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.665702] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.666953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e63274a-09b8-4231-84a8-599f52179c97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.678226] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.678644] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3110a19f-d022-4d06-bc95-fa688ad11323 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.686668] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 876.686668] env[62066]: value = "task-1156327" [ 876.686668] env[62066]: _type = "Task" [ 876.686668] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.695057] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156327, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.741562] env[62066]: DEBUG nova.network.neutron [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Updated VIF entry in instance network info cache for port fb8eff91-5d00-4cdc-87e0-8aa48d429972. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.742019] env[62066]: DEBUG nova.network.neutron [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Updating instance_info_cache with network_info: [{"id": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "address": "fa:16:3e:8f:59:4f", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb8eff91-5d", "ovs_interfaceid": "fb8eff91-5d00-4cdc-87e0-8aa48d429972", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.777540] env[62066]: DEBUG nova.compute.utils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 876.784035] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.784035] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.827111] env[62066]: DEBUG nova.policy [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1984a406905949b1b7653c4b235ccc23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b83b0c5ef67b4135b71299dfe2ec4263', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.008937] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.049136] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156324, 'name': CreateVM_Task, 'duration_secs': 1.511455} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.049136] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.050833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.050833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.050833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.051221] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccacef9e-cbc2-4078-87ba-922bceb71008 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.056389] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 877.056389] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5284bbe2-7954-14af-62a9-3b812353cca7" [ 877.056389] env[62066]: _type = "Task" [ 877.056389] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.064824] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5284bbe2-7954-14af-62a9-3b812353cca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.091825] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Successfully created port: d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.098886] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156326, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.136359] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27ad5e2b-cd20-47c3-8bbf-38739f6708d3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance 'ccb9f50f-dcc3-4d81-944e-d70803185ae1' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 877.160479] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.199912] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156327, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.244813] env[62066]: DEBUG oslo_concurrency.lockutils [req-615ca603-f9d3-4a6e-9e5d-c2c48ea57ec9 req-e2205420-2433-4398-8b82-aa27921cac29 service nova] Releasing lock "refresh_cache-0fda5f59-55ac-4150-8402-00064d14c8ab" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.282362] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.286202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e8ff9ad-8abb-4bc9-9998-520ea7280726 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.287363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 22.999s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.374210] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Updating instance_info_cache with network_info: [{"id": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "address": "fa:16:3e:02:f6:68", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55ac0005-05", "ovs_interfaceid": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.506549] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156325, 'name': ReconfigVM_Task, 'duration_secs': 0.704456} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.506910] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Reconfigured VM instance instance-0000004a to attach disk [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef/ecf9fd16-82c7-4bea-b6a9-7262e75effef.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.507244] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 877.571026] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5284bbe2-7954-14af-62a9-3b812353cca7, 'name': SearchDatastore_Task, 'duration_secs': 0.019573} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.571890] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.572223] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.572482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.572640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.572822] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.573162] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c105b177-aaa4-4f8f-bacf-19e9b18dc4fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.581445] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.581669] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.582389] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c4b940f-2ed7-4c9a-bfd3-4d008855f5ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.588370] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 877.588370] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5265b911-2d96-5329-3424-f7a06b12b901" [ 877.588370] env[62066]: _type = "Task" [ 877.588370] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.599602] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265b911-2d96-5329-3424-f7a06b12b901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.602832] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156326, 'name': ReconfigVM_Task, 'duration_secs': 0.63675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.603090] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 4914be5d-2cc8-48b7-96e5-9192e5c73fa3/4914be5d-2cc8-48b7-96e5-9192e5c73fa3.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.603699] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f0aa809-b9a2-49da-9cdc-7950ad9755e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.609691] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 877.609691] env[62066]: value = "task-1156328" [ 877.609691] env[62066]: _type = "Task" [ 877.609691] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.617170] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156328, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.698891] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156327, 'name': PowerOffVM_Task, 'duration_secs': 0.636221} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.699267] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.699450] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.699640] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaf213a1-6967-45b1-ada4-6a62f37a4dee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.780188] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.780188] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.780188] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Deleting the datastore file [datastore1] 3092f591-214f-40ba-a8e2-ccc6a0f007c9 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.780188] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b177458-141a-46a8-84a5-03955923ac7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.784630] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for the task: (returnval){ [ 877.784630] env[62066]: value = "task-1156330" [ 877.784630] env[62066]: _type = "Task" [ 877.784630] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.810216] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.859851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.860139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.860331] env[62066]: INFO nova.compute.manager [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Shelving [ 877.876816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.877443] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Instance network_info: |[{"id": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "address": "fa:16:3e:02:f6:68", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55ac0005-05", "ovs_interfaceid": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.877619] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:f6:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55ac0005-052f-46d1-9bf5-8807457b4fa8', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.885072] env[62066]: DEBUG oslo.service.loopingcall [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.885567] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.885745] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5035c90-1e56-46b4-852d-9bffd5b10a74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.905102] env[62066]: DEBUG nova.compute.manager [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Received event network-vif-plugged-55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.905325] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Acquiring lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.905522] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.905692] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.905860] env[62066]: DEBUG nova.compute.manager [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] No waiting events found dispatching network-vif-plugged-55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.906037] env[62066]: WARNING nova.compute.manager [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Received unexpected event network-vif-plugged-55ac0005-052f-46d1-9bf5-8807457b4fa8 for instance with vm_state building and task_state spawning. [ 877.906247] env[62066]: DEBUG nova.compute.manager [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Received event network-changed-55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.906448] env[62066]: DEBUG nova.compute.manager [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Refreshing instance network info cache due to event network-changed-55ac0005-052f-46d1-9bf5-8807457b4fa8. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.906653] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Acquiring lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.906798] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Acquired lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.906958] env[62066]: DEBUG nova.network.neutron [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Refreshing network info cache for port 55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.909758] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.909758] env[62066]: value = "task-1156331" [ 877.909758] env[62066]: _type = "Task" [ 877.909758] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.919465] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156331, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.014909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786b324c-a0f6-44dd-982b-084c999e5eb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.036811] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a95452-2f56-4ba4-96bd-1d4532345766 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.055839] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 878.103411] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265b911-2d96-5329-3424-f7a06b12b901, 'name': SearchDatastore_Task, 'duration_secs': 0.014252} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.104335] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7193b60a-1a30-4e8b-84c7-0a623063b17f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.109554] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 878.109554] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52957d1d-364f-e3d1-372a-631d9c445e4b" [ 878.109554] env[62066]: _type = "Task" [ 878.109554] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.121283] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52957d1d-364f-e3d1-372a-631d9c445e4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.124352] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156328, 'name': Rename_Task, 'duration_secs': 0.139708} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.124604] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.124841] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f097452-4da6-4cd3-ae93-8ab8ce1f47b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.131262] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 878.131262] env[62066]: value = "task-1156332" [ 878.131262] env[62066]: _type = "Task" [ 878.131262] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.139781] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156332, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.296674] env[62066]: DEBUG oslo_vmware.api [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Task: {'id': task-1156330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133691} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.296995] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.297244] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.297458] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.297667] env[62066]: INFO nova.compute.manager [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Took 1.63 seconds to destroy the instance on the hypervisor. [ 878.297947] env[62066]: DEBUG oslo.service.loopingcall [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.298230] env[62066]: DEBUG nova.compute.manager [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.298320] env[62066]: DEBUG nova.network.neutron [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 878.307225] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Applying migration context for instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 as it has an incoming, in-progress migration b2b4faf0-b29d-43d6-be2a-aeeeca362aa5. Migration status is finished {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 878.307662] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Applying migration context for instance ecf9fd16-82c7-4bea-b6a9-7262e75effef as it has an incoming, in-progress migration 45cece1f-bf1a-4a05-81c8-de8c7a858bdd. Migration status is migrating {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 878.309994] env[62066]: INFO nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating resource usage from migration b2b4faf0-b29d-43d6-be2a-aeeeca362aa5 [ 878.310654] env[62066]: INFO nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating resource usage from migration 45cece1f-bf1a-4a05-81c8-de8c7a858bdd [ 878.313135] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.336506] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.336774] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.336936] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.337134] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.337285] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.337438] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.337662] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.337851] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.338037] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.338208] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.338672] env[62066]: DEBUG nova.virt.hardware [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.339280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0736386-e9c7-4c9d-b05a-cda1dfc087c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.342909] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343062] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d50f0505-66f4-412f-9744-25e0ea96277a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343277] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a8dd7483-0588-4f60-9504-20de799e69f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343391] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance de9eb96d-e71b-4878-83b8-f75966dc3f48 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 878.343437] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343568] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343683] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 831c0bed-8a41-4672-aa08-d60b64f365b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343810] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 854a1ea4-515d-4a05-b179-22713b63f7c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.343937] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 68bf0e89-8236-44ef-b514-d25698eb23c3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 878.344062] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3092f591-214f-40ba-a8e2-ccc6a0f007c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.344194] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a2251e56-2787-412d-89c9-eef111ee6d2b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.344322] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 878.344450] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.344560] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f24f8067-07b2-4941-8464-c30b58983ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.344680] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 31e50f97-f873-44ad-9923-67923cdb8d3a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.344805] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance da8b3bca-dbf5-4266-826d-ed93772fa04c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 878.344938] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 878.345062] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 4914be5d-2cc8-48b7-96e5-9192e5c73fa3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.345174] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Migration b2b4faf0-b29d-43d6-be2a-aeeeca362aa5 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 878.345293] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.345399] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Migration 45cece1f-bf1a-4a05-81c8-de8c7a858bdd is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 878.345516] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ecf9fd16-82c7-4bea-b6a9-7262e75effef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.345621] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 0fda5f59-55ac-4150-8402-00064d14c8ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.345725] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f578eda3-3d81-418b-bcb7-0a954835ed72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.345842] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance cbf3c137-cc0f-42b7-96fb-2e1956e49b51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 878.346085] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 20 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 878.346231] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4480MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=20 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 878.353689] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c93614-bd27-41a8-b0aa-b5b572ced4df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.372236] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.374179] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afbfd7b3-d7bd-412b-905b-5d2a1c3e2a72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.380014] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 878.380014] env[62066]: value = "task-1156333" [ 878.380014] env[62066]: _type = "Task" [ 878.380014] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.387739] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156333, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.422392] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156331, 'name': CreateVM_Task, 'duration_secs': 0.476183} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.422561] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.423276] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.423441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.423819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.423999] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93f880e3-5cd7-43d2-8349-50b2f4e2b862 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.429721] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 878.429721] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52127472-3397-6ef1-1a29-f111d4ac8b98" [ 878.429721] env[62066]: _type = "Task" [ 878.429721] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.437072] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52127472-3397-6ef1-1a29-f111d4ac8b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.615490] env[62066]: DEBUG nova.network.neutron [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Port 3ea61126-db59-442a-91ed-a46b37e22462 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 878.628201] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52957d1d-364f-e3d1-372a-631d9c445e4b, 'name': SearchDatastore_Task, 'duration_secs': 0.011313} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.628934] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.629515] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 0fda5f59-55ac-4150-8402-00064d14c8ab/0fda5f59-55ac-4150-8402-00064d14c8ab.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.629909] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df4a3bef-f303-4ca1-8c0c-e1778c74f8cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.641536] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156332, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.643072] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 878.643072] env[62066]: value = "task-1156334" [ 878.643072] env[62066]: _type = "Task" [ 878.643072] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.658615] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156334, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.668840] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Successfully updated port: d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.745784] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd7f3f1-405d-4679-a8f6-12ada050bd1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.759576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b75c741-edb6-438e-a4d5-a1e9b72951c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.797237] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966a6988-9ee9-4673-a1c9-b3c601a0e382 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.805935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ba0a90-9a20-4261-8516-2f81a632d67c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.821045] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.891250] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156333, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.940535] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52127472-3397-6ef1-1a29-f111d4ac8b98, 'name': SearchDatastore_Task, 'duration_secs': 0.011275} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.940892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.941159] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.941412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.941611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.941895] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.942261] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7073f7ce-f1ae-445c-b74b-3e3a17a387d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.952916] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.953144] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.954474] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddf72d30-98f8-4d04-bc74-f9cc93ad78c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.960855] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 878.960855] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a47852-28b0-8e5d-0701-55f83ab382bc" [ 878.960855] env[62066]: _type = "Task" [ 878.960855] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.970033] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a47852-28b0-8e5d-0701-55f83ab382bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.999808] env[62066]: DEBUG nova.network.neutron [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Updated VIF entry in instance network info cache for port 55ac0005-052f-46d1-9bf5-8807457b4fa8. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.000576] env[62066]: DEBUG nova.network.neutron [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Updating instance_info_cache with network_info: [{"id": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "address": "fa:16:3e:02:f6:68", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55ac0005-05", "ovs_interfaceid": "55ac0005-052f-46d1-9bf5-8807457b4fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.144636] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156332, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.155985] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156334, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.172091] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.172238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.172418] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.325562] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.394992] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156333, 'name': PowerOffVM_Task, 'duration_secs': 0.586861} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.395997] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.397025] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2439ef3d-91f5-4349-a4ab-9770e87efc65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.416898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4757cbda-546d-458b-993d-bb50910c7e0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.427828] env[62066]: DEBUG nova.compute.manager [req-d82e3471-5dca-45de-adb5-8c9cfb1e3f97 req-bdfdc2a8-99d5-4fac-83e3-daf62cf9b14e service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Received event network-vif-deleted-f19a58d1-80e6-4f51-bd80-36927b0c3867 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.428048] env[62066]: INFO nova.compute.manager [req-d82e3471-5dca-45de-adb5-8c9cfb1e3f97 req-bdfdc2a8-99d5-4fac-83e3-daf62cf9b14e service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Neutron deleted interface f19a58d1-80e6-4f51-bd80-36927b0c3867; detaching it from the instance and deleting it from the info cache [ 879.428230] env[62066]: DEBUG nova.network.neutron [req-d82e3471-5dca-45de-adb5-8c9cfb1e3f97 req-bdfdc2a8-99d5-4fac-83e3-daf62cf9b14e service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.471425] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a47852-28b0-8e5d-0701-55f83ab382bc, 'name': SearchDatastore_Task, 'duration_secs': 0.05792} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.475225] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d6e8943-3f94-4067-8b6c-1536ed34cf3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.481504] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 879.481504] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529180c5-9d82-8f65-c808-413cf9460d56" [ 879.481504] env[62066]: _type = "Task" [ 879.481504] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.489137] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529180c5-9d82-8f65-c808-413cf9460d56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.503438] env[62066]: DEBUG oslo_concurrency.lockutils [req-3503176c-d69f-4c5b-af59-ff615a238b81 req-4bdb3c5d-1d68-485f-a211-7c24c2e58753 service nova] Releasing lock "refresh_cache-f578eda3-3d81-418b-bcb7-0a954835ed72" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.649351] env[62066]: DEBUG oslo_vmware.api [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156332, 'name': PowerOnVM_Task, 'duration_secs': 1.296048} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.655999] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.656272] env[62066]: INFO nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Took 16.19 seconds to spawn the instance on the hypervisor. [ 879.656496] env[62066]: DEBUG nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.660663] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.660932] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.661167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.662612] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4ad811-314f-4422-b3cc-307f0d2c578b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.672310] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156334, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540238} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.675188] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 0fda5f59-55ac-4150-8402-00064d14c8ab/0fda5f59-55ac-4150-8402-00064d14c8ab.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.675456] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.680569] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa10921e-5844-4d1d-b5ad-e8136ca60462 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.688438] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 879.688438] env[62066]: value = "task-1156335" [ 879.688438] env[62066]: _type = "Task" [ 879.688438] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.699108] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156335, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.710813] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.831257] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 879.835381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.548s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.835708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.523s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.835909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.838085] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.549s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.838363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.840172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.817s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.840320] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.842924] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.906s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.843184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.844712] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.689s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.844914] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.852048] env[62066]: DEBUG nova.network.neutron [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.876198] env[62066]: INFO nova.scheduler.client.report [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d [ 879.879917] env[62066]: INFO nova.scheduler.client.report [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Deleted allocations for instance da8b3bca-dbf5-4266-826d-ed93772fa04c [ 879.894836] env[62066]: INFO nova.scheduler.client.report [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocations for instance 68bf0e89-8236-44ef-b514-d25698eb23c3 [ 879.901461] env[62066]: INFO nova.scheduler.client.report [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Deleted allocations for instance de9eb96d-e71b-4878-83b8-f75966dc3f48 [ 879.921382] env[62066]: INFO nova.scheduler.client.report [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15 [ 879.930730] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 879.931330] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-61636f6d-7f30-4926-9ecf-e789fc2cf2a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.934818] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26f571d8-1a45-4197-914c-37a77e837699 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.943021] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 879.943021] env[62066]: value = "task-1156336" [ 879.943021] env[62066]: _type = "Task" [ 879.943021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.947336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f13edc-2486-48df-a606-7c7b42c341e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.968489] env[62066]: DEBUG nova.network.neutron [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Updating instance_info_cache with network_info: [{"id": "d9083d7c-317d-46e6-bf0f-eb924b695754", "address": "fa:16:3e:95:ca:af", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9083d7c-31", "ovs_interfaceid": "d9083d7c-317d-46e6-bf0f-eb924b695754", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.977562] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156336, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.978979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.979238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.979485] env[62066]: DEBUG nova.compute.manager [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Going to confirm migration 2 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 879.982709] env[62066]: DEBUG nova.compute.manager [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Received event network-vif-plugged-d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.982709] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Acquiring lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.982709] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.982916] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.983087] env[62066]: DEBUG nova.compute.manager [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] No waiting events found dispatching network-vif-plugged-d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.983268] env[62066]: WARNING nova.compute.manager [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Received unexpected event network-vif-plugged-d9083d7c-317d-46e6-bf0f-eb924b695754 for instance with vm_state building and task_state spawning. [ 879.983428] env[62066]: DEBUG nova.compute.manager [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Received event network-changed-d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.983591] env[62066]: DEBUG nova.compute.manager [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Refreshing instance network info cache due to event network-changed-d9083d7c-317d-46e6-bf0f-eb924b695754. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.983808] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Acquiring lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.999261] env[62066]: DEBUG nova.compute.manager [req-d82e3471-5dca-45de-adb5-8c9cfb1e3f97 req-bdfdc2a8-99d5-4fac-83e3-daf62cf9b14e service nova] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Detach interface failed, port_id=f19a58d1-80e6-4f51-bd80-36927b0c3867, reason: Instance 3092f591-214f-40ba-a8e2-ccc6a0f007c9 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 880.009758] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529180c5-9d82-8f65-c808-413cf9460d56, 'name': SearchDatastore_Task, 'duration_secs': 0.018545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.009758] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.010044] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f578eda3-3d81-418b-bcb7-0a954835ed72/f578eda3-3d81-418b-bcb7-0a954835ed72.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.010367] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d25ce9d7-9e80-4d90-af0f-4a380722e485 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.016785] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 880.016785] env[62066]: value = "task-1156337" [ 880.016785] env[62066]: _type = "Task" [ 880.016785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.027892] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.188688] env[62066]: INFO nova.compute.manager [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Took 43.75 seconds to build instance. [ 880.195595] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 880.204735] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156335, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06205} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.206742] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.206742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09d4dba-c480-4885-aa8f-dfa3e04dfc51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.242077] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 0fda5f59-55ac-4150-8402-00064d14c8ab/0fda5f59-55ac-4150-8402-00064d14c8ab.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.242997] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fa6fbdc-a978-4943-b544-73416084a73b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.264494] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 880.264494] env[62066]: value = "task-1156338" [ 880.264494] env[62066]: _type = "Task" [ 880.264494] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.273897] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156338, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.354300] env[62066]: INFO nova.compute.manager [-] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Took 2.06 seconds to deallocate network for instance. [ 880.389531] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c957e2ec-3aed-4b4e-870e-028ccd331b6b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3b712d48-1365-4eed-8d6d-8ee6dbf51c2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.853s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.391496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d761890d-00e5-48e1-8db5-052a908fc239 tempest-InstanceActionsNegativeTestJSON-872525787 tempest-InstanceActionsNegativeTestJSON-872525787-project-member] Lock "da8b3bca-dbf5-4266-826d-ed93772fa04c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.926s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.410748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c79573b-f918-47b1-8790-b4b566ff2441 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "68bf0e89-8236-44ef-b514-d25698eb23c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.227s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.412917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5f422857-7ae9-4e98-8a00-69bb95752c03 tempest-ServerRescueTestJSON-768309486 tempest-ServerRescueTestJSON-768309486-project-member] Lock "de9eb96d-e71b-4878-83b8-f75966dc3f48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.374s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.432550] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bf84cbc2-3443-4d7b-a1af-bb1cbd3bd230 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "6bd9da25-fcfc-41a1-b5d6-07d28ab37e15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.096s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.454426] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156336, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.473330] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.473871] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Instance network_info: |[{"id": "d9083d7c-317d-46e6-bf0f-eb924b695754", "address": "fa:16:3e:95:ca:af", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9083d7c-31", "ovs_interfaceid": "d9083d7c-317d-46e6-bf0f-eb924b695754", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.474309] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Acquired lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.474584] env[62066]: DEBUG nova.network.neutron [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Refreshing network info cache for port d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.476094] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:ca:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9083d7c-317d-46e6-bf0f-eb924b695754', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.484492] env[62066]: DEBUG oslo.service.loopingcall [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.488503] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.489123] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02fb5ff3-6e4b-45e5-9f6b-444425b94e0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.515546] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.515546] env[62066]: value = "task-1156339" [ 880.515546] env[62066]: _type = "Task" [ 880.515546] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.525575] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156339, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.531054] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156337, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.565523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.640066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.640254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.640470] env[62066]: DEBUG nova.network.neutron [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.640718] env[62066]: DEBUG nova.objects.instance [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'info_cache' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.690879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1e2b52-a306-4be7-800f-7d702b883dfc tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.271s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.691130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.126s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.691349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.691561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.691775] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.693790] env[62066]: INFO nova.compute.manager [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Terminating instance [ 880.695558] env[62066]: DEBUG nova.compute.manager [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.695753] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.696606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb085e81-1690-4853-9b16-8d88f00b51df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.709691] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.709944] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d245066-35a8-4e55-8569-61decb56c262 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.719214] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 880.719214] env[62066]: value = "task-1156340" [ 880.719214] env[62066]: _type = "Task" [ 880.719214] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.730192] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156340, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.749429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.749625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.749968] env[62066]: DEBUG nova.network.neutron [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.774967] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156338, 'name': ReconfigVM_Task, 'duration_secs': 0.370656} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.775294] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 0fda5f59-55ac-4150-8402-00064d14c8ab/0fda5f59-55ac-4150-8402-00064d14c8ab.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.775989] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d61a9577-8d34-41a8-be95-33b871d98dc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.782572] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 880.782572] env[62066]: value = "task-1156341" [ 880.782572] env[62066]: _type = "Task" [ 880.782572] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.796067] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156341, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.862359] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.863021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.863157] env[62066]: DEBUG nova.objects.instance [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lazy-loading 'resources' on Instance uuid 3092f591-214f-40ba-a8e2-ccc6a0f007c9 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.952682] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156336, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.030517] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156339, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.035261] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.035763] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f578eda3-3d81-418b-bcb7-0a954835ed72/f578eda3-3d81-418b-bcb7-0a954835ed72.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.036098] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.036386] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ebd07ed8-2953-4f6e-8c93-4a1cb566d412 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.044379] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 881.044379] env[62066]: value = "task-1156342" [ 881.044379] env[62066]: _type = "Task" [ 881.044379] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.051973] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.186731] env[62066]: DEBUG nova.network.neutron [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Updated VIF entry in instance network info cache for port d9083d7c-317d-46e6-bf0f-eb924b695754. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.187137] env[62066]: DEBUG nova.network.neutron [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Updating instance_info_cache with network_info: [{"id": "d9083d7c-317d-46e6-bf0f-eb924b695754", "address": "fa:16:3e:95:ca:af", "network": {"id": "c3007381-c473-45ce-9283-b278ebbcb42e", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-821733123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b83b0c5ef67b4135b71299dfe2ec4263", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9083d7c-31", "ovs_interfaceid": "d9083d7c-317d-46e6-bf0f-eb924b695754", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.230956] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156340, 'name': PowerOffVM_Task, 'duration_secs': 0.25125} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.231331] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.232792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.232792] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c70d6c01-02f5-4c25-aa9b-bf7a8db8d9f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.300236] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156341, 'name': Rename_Task, 'duration_secs': 0.146276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.300793] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.301245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd0569ff-8e73-4c52-b5b8-e71d819013e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.306076] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.306467] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.306867] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleting the datastore file [datastore1] 4914be5d-2cc8-48b7-96e5-9192e5c73fa3 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.307769] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ab0e373-e3d5-47ea-b842-40d2ab506eef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.315703] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 881.315703] env[62066]: value = "task-1156344" [ 881.315703] env[62066]: _type = "Task" [ 881.315703] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.317828] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 881.317828] env[62066]: value = "task-1156345" [ 881.317828] env[62066]: _type = "Task" [ 881.317828] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.334865] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156344, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.340006] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156345, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.459411] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156336, 'name': CreateSnapshot_Task, 'duration_secs': 1.036493} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.459701] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 881.460509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac3df13-8af9-4266-82f8-e60f65ce0d2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.526952] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156339, 'name': CreateVM_Task, 'duration_secs': 0.718729} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.530266] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.535216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.535300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.535627] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.540020] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bc23783-b06a-475a-93a7-81165acb99bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.542972] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 881.542972] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5210d840-c9b4-8aed-c96e-883f7982b443" [ 881.542972] env[62066]: _type = "Task" [ 881.542972] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.559406] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5210d840-c9b4-8aed-c96e-883f7982b443, 'name': SearchDatastore_Task, 'duration_secs': 0.008871} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.562431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.562683] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.562920] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.563086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.563280] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.563568] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.216791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.566123] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-641fc118-b122-46a7-b206-c54f46b59a7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.568007] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.568953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084ed9b6-890d-4405-be38-36714682b1ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.591988] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] f578eda3-3d81-418b-bcb7-0a954835ed72/f578eda3-3d81-418b-bcb7-0a954835ed72.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.595550] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c6ef4a8-0220-44e8-a0ed-8acbd6b9adab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.609791] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.609976] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.611192] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b84b8c8-e515-4e29-93de-9d94515fe083 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.619023] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 881.619023] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525fd253-df9c-ac6c-ba59-7232f4699fe2" [ 881.619023] env[62066]: _type = "Task" [ 881.619023] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.626022] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 881.626022] env[62066]: value = "task-1156346" [ 881.626022] env[62066]: _type = "Task" [ 881.626022] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.629122] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525fd253-df9c-ac6c-ba59-7232f4699fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.636702] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.689597] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac5a55f6-e474-443b-9786-110a07a271ec req-16495a9c-35ca-4106-82a3-946f822d4c5c service nova] Releasing lock "refresh_cache-cbf3c137-cc0f-42b7-96fb-2e1956e49b51" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.766453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca8cf44-7ad7-4432-a09d-bab42db4b085 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.774254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867800e9-3e51-495d-b077-816ce0a2285f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.812370] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5759a5e-fbc1-4530-b163-f7b5a10c8778 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.823448] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef1a3d6-b91b-4b43-a9dc-01850ab4940d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.834280] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156344, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.843059] env[62066]: DEBUG nova.compute.provider_tree [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.847287] env[62066]: DEBUG oslo_vmware.api [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156345, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137494} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.847837] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.848086] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.848294] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.848480] env[62066]: INFO nova.compute.manager [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 881.848749] env[62066]: DEBUG oslo.service.loopingcall [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.848964] env[62066]: DEBUG nova.compute.manager [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.849072] env[62066]: DEBUG nova.network.neutron [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 881.900782] env[62066]: DEBUG nova.network.neutron [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.916561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "b2afea5a-2390-40b0-b860-c50f71ebac94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.916810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.981453] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 881.981856] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2342ec5a-37c4-4af5-998e-7b709477bff4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.990895] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 881.990895] env[62066]: value = "task-1156347" [ 881.990895] env[62066]: _type = "Task" [ 881.990895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.999376] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156347, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.136626] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525fd253-df9c-ac6c-ba59-7232f4699fe2, 'name': SearchDatastore_Task, 'duration_secs': 0.010431} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.142252] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.142542] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-601545ac-e58a-4e51-8dfe-da3fdacbd718 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.151392] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 882.151392] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52877220-2d6b-7bb3-bbd2-7f4b3a3c2426" [ 882.151392] env[62066]: _type = "Task" [ 882.151392] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.170694] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52877220-2d6b-7bb3-bbd2-7f4b3a3c2426, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.330851] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156344, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.349541] env[62066]: DEBUG nova.scheduler.client.report [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.403510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.419532] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 882.435685] env[62066]: DEBUG nova.network.neutron [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.504927] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156347, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.548824] env[62066]: DEBUG nova.compute.manager [req-6916722a-a071-496f-b329-accb84df66e1 req-cc40b84e-97b0-4a8c-bfdb-65675049d3bd service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Received event network-vif-deleted-dc192027-6e82-44a5-8f5c-95cbc8ea621c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.548869] env[62066]: INFO nova.compute.manager [req-6916722a-a071-496f-b329-accb84df66e1 req-cc40b84e-97b0-4a8c-bfdb-65675049d3bd service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Neutron deleted interface dc192027-6e82-44a5-8f5c-95cbc8ea621c; detaching it from the instance and deleting it from the info cache [ 882.549560] env[62066]: DEBUG nova.network.neutron [req-6916722a-a071-496f-b329-accb84df66e1 req-cc40b84e-97b0-4a8c-bfdb-65675049d3bd service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.633464] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156346, 'name': ReconfigVM_Task, 'duration_secs': 0.927374} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.634061] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Reconfigured VM instance instance-00000053 to attach disk [datastore1] f578eda3-3d81-418b-bcb7-0a954835ed72/f578eda3-3d81-418b-bcb7-0a954835ed72.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.634450] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d3f4d16-9e20-41de-8c6e-8388c7bc9b4f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.641584] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 882.641584] env[62066]: value = "task-1156348" [ 882.641584] env[62066]: _type = "Task" [ 882.641584] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.653859] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156348, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.669828] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52877220-2d6b-7bb3-bbd2-7f4b3a3c2426, 'name': SearchDatastore_Task, 'duration_secs': 0.030916} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.669921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.672020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] cbf3c137-cc0f-42b7-96fb-2e1956e49b51/cbf3c137-cc0f-42b7-96fb-2e1956e49b51.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.672020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2f725f4-467a-4b8e-973e-65e57ee812fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.680234] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 882.680234] env[62066]: value = "task-1156349" [ 882.680234] env[62066]: _type = "Task" [ 882.680234] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.695022] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.801830] env[62066]: DEBUG nova.network.neutron [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.833721] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156344, 'name': PowerOnVM_Task, 'duration_secs': 1.039597} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.834014] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.834250] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Took 9.80 seconds to spawn the instance on the hypervisor. [ 882.834465] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.835362] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dd6ee8-d659-44b2-a6ec-349f4797554a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.857720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.994s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.887298] env[62066]: INFO nova.scheduler.client.report [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Deleted allocations for instance 3092f591-214f-40ba-a8e2-ccc6a0f007c9 [ 882.936932] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca8e520-2a5f-456e-9e05-b46b62ea0a37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.945031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.946131] env[62066]: DEBUG nova.objects.instance [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'migration_context' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.949967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.951645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.952772] env[62066]: INFO nova.compute.claims [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.981332] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8916c3-604d-48a0-8ac9-90dd1ad440ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.993599] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 883.009161] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156347, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.055191] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-743266d1-7560-4460-9aee-6ee35cc7c0fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.061903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33016e9-dac6-483b-9742-924371f1e475 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.101201] env[62066]: DEBUG nova.compute.manager [req-6916722a-a071-496f-b329-accb84df66e1 req-cc40b84e-97b0-4a8c-bfdb-65675049d3bd service nova] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Detach interface failed, port_id=dc192027-6e82-44a5-8f5c-95cbc8ea621c, reason: Instance 4914be5d-2cc8-48b7-96e5-9192e5c73fa3 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 883.153555] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156348, 'name': Rename_Task, 'duration_secs': 0.134835} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.153932] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.154359] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0bfa0d3e-e917-4709-853e-07bd21387019 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.161122] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 883.161122] env[62066]: value = "task-1156350" [ 883.161122] env[62066]: _type = "Task" [ 883.161122] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.169622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.169933] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.177068] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156350, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.193479] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506979} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.194036] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] cbf3c137-cc0f-42b7-96fb-2e1956e49b51/cbf3c137-cc0f-42b7-96fb-2e1956e49b51.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.194614] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.194614] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6ac2667-e6f7-478d-8282-64258ac37f46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.203426] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 883.203426] env[62066]: value = "task-1156351" [ 883.203426] env[62066]: _type = "Task" [ 883.203426] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.219353] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.307083] env[62066]: INFO nova.compute.manager [-] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Took 1.46 seconds to deallocate network for instance. [ 883.359033] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Took 39.51 seconds to build instance. [ 883.398721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d59b2081-1f8e-42fe-8e21-5a8ec733e780 tempest-ServersTestJSON-109668587 tempest-ServersTestJSON-109668587-project-member] Lock "3092f591-214f-40ba-a8e2-ccc6a0f007c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.741s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.449912] env[62066]: DEBUG nova.objects.base [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 883.452290] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87975799-d9c4-40a0-bf47-b8d2a644bc3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.479577] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a838e6a2-dbac-4432-b266-d04f24d9c4b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.488688] env[62066]: DEBUG oslo_vmware.api [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 883.488688] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529c2554-0043-97ac-bcbf-06c738148376" [ 883.488688] env[62066]: _type = "Task" [ 883.488688] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.502050] env[62066]: DEBUG oslo_vmware.api [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529c2554-0043-97ac-bcbf-06c738148376, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.506158] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.510943] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48243999-270b-4387-92b4-e4e7fc17fafe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.513079] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156347, 'name': CloneVM_Task, 'duration_secs': 1.0987} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.513304] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Created linked-clone VM from snapshot [ 883.514447] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9283f5f1-2ea6-4ffa-ab05-83f5ba7962c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.519125] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 883.519125] env[62066]: value = "task-1156352" [ 883.519125] env[62066]: _type = "Task" [ 883.519125] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.526258] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Uploading image c96c784e-9ccb-46d7-b666-d95d5fb00a8b {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 883.532733] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.553575] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 883.553575] env[62066]: value = "vm-251727" [ 883.553575] env[62066]: _type = "VirtualMachine" [ 883.553575] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 883.554241] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b9ccf5a3-85a5-4877-bc48-0acaad2d9ed5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.561327] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lease: (returnval){ [ 883.561327] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52571635-0e98-c062-e067-6118c3e29863" [ 883.561327] env[62066]: _type = "HttpNfcLease" [ 883.561327] env[62066]: } obtained for exporting VM: (result){ [ 883.561327] env[62066]: value = "vm-251727" [ 883.561327] env[62066]: _type = "VirtualMachine" [ 883.561327] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 883.561614] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the lease: (returnval){ [ 883.561614] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52571635-0e98-c062-e067-6118c3e29863" [ 883.561614] env[62066]: _type = "HttpNfcLease" [ 883.561614] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.569632] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.569632] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52571635-0e98-c062-e067-6118c3e29863" [ 883.569632] env[62066]: _type = "HttpNfcLease" [ 883.569632] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.673544] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.680123] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156350, 'name': PowerOnVM_Task, 'duration_secs': 0.457583} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.680123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.680123] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Took 8.03 seconds to spawn the instance on the hypervisor. [ 883.680123] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.680123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6aa7424-6001-42ed-b556-a56d168a6bbe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.726037] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06529} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.726151] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.727388] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f253bbda-249a-4b15-9131-6b1ad87e842b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.768077] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] cbf3c137-cc0f-42b7-96fb-2e1956e49b51/cbf3c137-cc0f-42b7-96fb-2e1956e49b51.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.769023] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c027ff1-d534-45d3-a70b-4d1c5d532fed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.800791] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 883.800791] env[62066]: value = "task-1156354" [ 883.800791] env[62066]: _type = "Task" [ 883.800791] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.813239] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156354, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.814529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.861217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.035s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.003617] env[62066]: DEBUG oslo_vmware.api [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529c2554-0043-97ac-bcbf-06c738148376, 'name': SearchDatastore_Task, 'duration_secs': 0.00941} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.003617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.033613] env[62066]: DEBUG oslo_vmware.api [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156352, 'name': PowerOnVM_Task, 'duration_secs': 0.439752} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.033902] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.034107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fd87a023-40d4-467d-b709-b9e182a4965c tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance 'ecf9fd16-82c7-4bea-b6a9-7262e75effef' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 884.071755] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.071755] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52571635-0e98-c062-e067-6118c3e29863" [ 884.071755] env[62066]: _type = "HttpNfcLease" [ 884.071755] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 884.072308] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 884.072308] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52571635-0e98-c062-e067-6118c3e29863" [ 884.072308] env[62066]: _type = "HttpNfcLease" [ 884.072308] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 884.072956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab731e7-8421-46e3-8337-b84244acf4d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.084541] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 884.084541] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 884.197127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.199778] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Took 40.33 seconds to build instance. [ 884.208049] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f8abf88b-34c2-49b7-8d16-25db6445e217 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.311804] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156354, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.402027] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d3241c-8f41-48ec-ba94-90a54395597e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.409387] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55816013-e814-4bc5-a194-54c3fd637b15 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.448038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28dd67e3-a09e-46d2-8c9c-4baff744d2ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.456686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76729bda-4230-4580-a267-e5b2df817a84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.469948] env[62066]: DEBUG nova.compute.provider_tree [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.701995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.840s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.812201] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156354, 'name': ReconfigVM_Task, 'duration_secs': 0.74794} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.814374] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Reconfigured VM instance instance-00000054 to attach disk [datastore1] cbf3c137-cc0f-42b7-96fb-2e1956e49b51/cbf3c137-cc0f-42b7-96fb-2e1956e49b51.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.814374] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5331a1ae-fb08-422b-92fe-46293dba9e45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.825025] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 884.825025] env[62066]: value = "task-1156355" [ 884.825025] env[62066]: _type = "Task" [ 884.825025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.833078] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156355, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.980670] env[62066]: DEBUG nova.scheduler.client.report [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.045799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.046198] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.334609] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156355, 'name': Rename_Task, 'duration_secs': 0.203489} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.335065] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.335418] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab1a7471-fa30-4ab9-b7b9-98c3175fffcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.349901] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 885.349901] env[62066]: value = "task-1156356" [ 885.349901] env[62066]: _type = "Task" [ 885.349901] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.362284] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.488996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.489679] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.497186] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.681s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.497482] env[62066]: DEBUG nova.objects.instance [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lazy-loading 'resources' on Instance uuid 4914be5d-2cc8-48b7-96e5-9192e5c73fa3 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.548472] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.699512] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 885.861486] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156356, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.003073] env[62066]: DEBUG nova.compute.utils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 886.007777] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 886.007983] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 886.059017] env[62066]: DEBUG nova.policy [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.074643] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.207468] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.361435] env[62066]: DEBUG oslo_vmware.api [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156356, 'name': PowerOnVM_Task, 'duration_secs': 0.729244} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.361948] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.362146] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Took 8.05 seconds to spawn the instance on the hypervisor. [ 886.362422] env[62066]: DEBUG nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.365975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8411a446-8aeb-45d0-a83e-4aef971c6277 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.397344] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48500e07-9bce-40c4-8748-89e2410433fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.403736] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87fbdbf-bf6e-4439-8319-96b9cec560c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.443860] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be8d45e-8331-478b-8906-7d3171fb6825 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.459273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01217411-df44-46bb-95d2-ba14ea529ca3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.472448] env[62066]: DEBUG nova.compute.provider_tree [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.514510] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.632323] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Successfully created port: cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.887976] env[62066]: INFO nova.compute.manager [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Took 42.98 seconds to build instance. [ 886.897850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.898237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.898503] env[62066]: DEBUG nova.compute.manager [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Going to confirm migration 3 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 886.976115] env[62066]: DEBUG nova.scheduler.client.report [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.390386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-53f2068b-19e9-4acd-a532-3c849334714d tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.492s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.483041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.485707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.483s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.511236] env[62066]: INFO nova.scheduler.client.report [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted allocations for instance 4914be5d-2cc8-48b7-96e5-9192e5c73fa3 [ 887.526061] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.539628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.539822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.540042] env[62066]: DEBUG nova.network.neutron [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.540237] env[62066]: DEBUG nova.objects.instance [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'info_cache' on Instance uuid ecf9fd16-82c7-4bea-b6a9-7262e75effef {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.567211] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.567479] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.567663] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.567884] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.567991] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.568184] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.568400] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.568563] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.568735] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.568907] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.569091] env[62066]: DEBUG nova.virt.hardware [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.573214] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4153a0b-71c6-421b-a935-95aecc380755 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.582181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47d1d17-532e-484e-b0a0-d9773ee35cd8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.024425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a54b8ab6-174d-4ac7-adbb-3bcd0da00b6f tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "4914be5d-2cc8-48b7-96e5-9192e5c73fa3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.333s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.380998] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12d819c-227f-488b-9e18-142639124226 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.391590] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcac99f3-ee7a-4558-bbfd-bb01699301cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.398343] env[62066]: DEBUG nova.compute.manager [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Received event network-vif-plugged-cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.398343] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] Acquiring lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.398343] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.398343] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.398525] env[62066]: DEBUG nova.compute.manager [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] No waiting events found dispatching network-vif-plugged-cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.398603] env[62066]: WARNING nova.compute.manager [req-5d1d8d39-49c0-4775-ba9e-3763347004bd req-bab984a2-7cc9-48a8-8152-6b2c9fce9289 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Received unexpected event network-vif-plugged-cc490fbd-bf0f-41fd-a20a-8438810bab18 for instance with vm_state building and task_state spawning. [ 888.444037] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9639bd14-6f54-4cde-a8d8-d67f2b7b1b25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.453337] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0575190c-f4bf-4995-8cb6-9f3a04052f48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.469570] env[62066]: DEBUG nova.compute.provider_tree [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.506247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "0fda5f59-55ac-4150-8402-00064d14c8ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.506629] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.506779] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.507048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.507185] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.509702] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Successfully updated port: cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.514285] env[62066]: INFO nova.compute.manager [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Terminating instance [ 888.514285] env[62066]: DEBUG nova.compute.manager [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.514285] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.515496] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f3f253-9d1b-4130-8026-c9db10c12e9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.524510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.524795] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80aedca7-66c0-4ff3-b137-d2f3ad49506e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.531627] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 888.531627] env[62066]: value = "task-1156357" [ 888.531627] env[62066]: _type = "Task" [ 888.531627] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.540559] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156357, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.947840] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "854a1ea4-515d-4a05-b179-22713b63f7c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.947840] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.948085] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.948228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.948473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.954476] env[62066]: INFO nova.compute.manager [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Terminating instance [ 888.963056] env[62066]: DEBUG nova.compute.manager [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.963435] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.964476] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca15eae-098a-44a6-a611-77c57b9fce2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.974691] env[62066]: DEBUG nova.scheduler.client.report [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.978571] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.981135] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1695096f-bb16-4b6f-93d3-637584ce4e12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.989045] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 888.989045] env[62066]: value = "task-1156358" [ 888.989045] env[62066]: _type = "Task" [ 888.989045] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.000472] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.015060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.015060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.015060] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.049583] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156357, 'name': PowerOffVM_Task, 'duration_secs': 0.237598} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.050978] env[62066]: DEBUG nova.network.neutron [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [{"id": "3ea61126-db59-442a-91ed-a46b37e22462", "address": "fa:16:3e:6f:9a:40", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea61126-db", "ovs_interfaceid": "3ea61126-db59-442a-91ed-a46b37e22462", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.052537] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.053184] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.054461] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bac6015-af08-4ab8-b5c7-f9117ef138c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.160721] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.160960] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.161154] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleting the datastore file [datastore1] 0fda5f59-55ac-4150-8402-00064d14c8ab {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.161419] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86b63f2e-1d85-48e0-a0fa-55d1eb75cf54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.174323] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 889.174323] env[62066]: value = "task-1156360" [ 889.174323] env[62066]: _type = "Task" [ 889.174323] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.185397] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.503335] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156358, 'name': PowerOffVM_Task, 'duration_secs': 0.351611} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.503335] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.503335] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.503335] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc7a8a31-05b1-4ac8-81bf-fa8776815823 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.554291] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-ecf9fd16-82c7-4bea-b6a9-7262e75effef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.554596] env[62066]: DEBUG nova.objects.instance [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'migration_context' on Instance uuid ecf9fd16-82c7-4bea-b6a9-7262e75effef {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.577415] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.577729] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.577969] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleting the datastore file [datastore1] 854a1ea4-515d-4a05-b179-22713b63f7c3 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.578897] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.581661] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18b55afb-c587-4930-84ae-096635204182 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.588797] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 889.588797] env[62066]: value = "task-1156362" [ 889.588797] env[62066]: _type = "Task" [ 889.588797] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.599182] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.687232] env[62066]: DEBUG oslo_vmware.api [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156360, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.292027} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.687707] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.687942] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.688121] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.688320] env[62066]: INFO nova.compute.manager [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Took 1.17 seconds to destroy the instance on the hypervisor. [ 889.688569] env[62066]: DEBUG oslo.service.loopingcall [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.691933] env[62066]: DEBUG nova.compute.manager [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.691991] env[62066]: DEBUG nova.network.neutron [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 889.783445] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.783773] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.829299] env[62066]: DEBUG nova.network.neutron [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Updating instance_info_cache with network_info: [{"id": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "address": "fa:16:3e:76:57:87", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc490fbd-bf", "ovs_interfaceid": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.991572] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.991572] env[62066]: DEBUG nova.compute.manager [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62066) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 889.996165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.799s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.999336] env[62066]: INFO nova.compute.claims [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.038392] env[62066]: DEBUG nova.compute.manager [req-fbfd249f-19d5-4d1d-ba12-db7e3f4c0044 req-bd697a83-a99b-4faf-b144-812ff6b760eb service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Received event network-vif-deleted-fb8eff91-5d00-4cdc-87e0-8aa48d429972 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.041263] env[62066]: INFO nova.compute.manager [req-fbfd249f-19d5-4d1d-ba12-db7e3f4c0044 req-bd697a83-a99b-4faf-b144-812ff6b760eb service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Neutron deleted interface fb8eff91-5d00-4cdc-87e0-8aa48d429972; detaching it from the instance and deleting it from the info cache [ 890.041511] env[62066]: DEBUG nova.network.neutron [req-fbfd249f-19d5-4d1d-ba12-db7e3f4c0044 req-bd697a83-a99b-4faf-b144-812ff6b760eb service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.057846] env[62066]: DEBUG nova.objects.base [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 890.059157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b954359-48d2-4f48-aefe-5e49bb0e153a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.083667] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02c32a61-4477-48a4-b4ed-e633f98bffa1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.091039] env[62066]: DEBUG oslo_vmware.api [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 890.091039] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525584ce-82d3-5af2-071b-835400a7fbdc" [ 890.091039] env[62066]: _type = "Task" [ 890.091039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.106890] env[62066]: DEBUG oslo_vmware.api [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.410644} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.111352] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.111956] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.112256] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.112668] env[62066]: INFO nova.compute.manager [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Took 1.15 seconds to destroy the instance on the hypervisor. [ 890.112948] env[62066]: DEBUG oslo.service.loopingcall [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.113306] env[62066]: DEBUG oslo_vmware.api [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525584ce-82d3-5af2-071b-835400a7fbdc, 'name': SearchDatastore_Task, 'duration_secs': 0.007791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.113556] env[62066]: DEBUG nova.compute.manager [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.113687] env[62066]: DEBUG nova.network.neutron [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.116403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.286934] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.334255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.334710] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Instance network_info: |[{"id": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "address": "fa:16:3e:76:57:87", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc490fbd-bf", "ovs_interfaceid": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.335521] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:57:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc490fbd-bf0f-41fd-a20a-8438810bab18', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.346197] env[62066]: DEBUG oslo.service.loopingcall [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.346197] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.346197] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-556d89eb-abdb-4000-a39d-b9fd7d07dc9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.367858] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.367858] env[62066]: value = "task-1156363" [ 890.367858] env[62066]: _type = "Task" [ 890.367858] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.379454] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156363, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.425164] env[62066]: DEBUG nova.compute.manager [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Received event network-changed-cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.425359] env[62066]: DEBUG nova.compute.manager [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Refreshing instance network info cache due to event network-changed-cc490fbd-bf0f-41fd-a20a-8438810bab18. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.425666] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] Acquiring lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.425846] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] Acquired lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.426266] env[62066]: DEBUG nova.network.neutron [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Refreshing network info cache for port cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.509582] env[62066]: DEBUG nova.network.neutron [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.544923] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c935992-d01e-4dd7-aced-03c32e93fabf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.554709] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead160e5-e139-415e-b239-4ceca196f111 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.567668] env[62066]: INFO nova.scheduler.client.report [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted allocation for migration b2b4faf0-b29d-43d6-be2a-aeeeca362aa5 [ 890.592692] env[62066]: DEBUG nova.compute.manager [req-fbfd249f-19d5-4d1d-ba12-db7e3f4c0044 req-bd697a83-a99b-4faf-b144-812ff6b760eb service nova] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Detach interface failed, port_id=fb8eff91-5d00-4cdc-87e0-8aa48d429972, reason: Instance 0fda5f59-55ac-4150-8402-00064d14c8ab could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 890.820978] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.882308] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156363, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.893407] env[62066]: DEBUG nova.network.neutron [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.014621] env[62066]: INFO nova.compute.manager [-] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Took 1.32 seconds to deallocate network for instance. [ 891.074021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-998ebcf0-bf3e-47cc-a2e9-39fdd3425502 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.095s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.196831] env[62066]: DEBUG nova.network.neutron [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Updated VIF entry in instance network info cache for port cc490fbd-bf0f-41fd-a20a-8438810bab18. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.197284] env[62066]: DEBUG nova.network.neutron [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Updating instance_info_cache with network_info: [{"id": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "address": "fa:16:3e:76:57:87", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc490fbd-bf", "ovs_interfaceid": "cc490fbd-bf0f-41fd-a20a-8438810bab18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.378675] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156363, 'name': CreateVM_Task, 'duration_secs': 0.543567} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.379759] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.380656] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468ca79d-54f8-46e5-b1f1-7a197808ef54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.383982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.384346] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.384718] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.385441] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9a4a974-05d9-4622-a061-3e412a166ae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.393659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f11210-9808-4b22-9c98-d203c8082a44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.397320] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 891.397320] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52dc8a31-c162-01db-d222-4de9ec469f4b" [ 891.397320] env[62066]: _type = "Task" [ 891.397320] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.397747] env[62066]: INFO nova.compute.manager [-] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Took 1.28 seconds to deallocate network for instance. [ 891.442093] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bee8f48-6a60-431a-b38c-0bb02d54c3df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.448841] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52dc8a31-c162-01db-d222-4de9ec469f4b, 'name': SearchDatastore_Task, 'duration_secs': 0.015582} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.449650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.449973] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.450280] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.450442] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.450655] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.451328] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-833035e8-4fb3-4ac5-b979-bc6f7a3b4a19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.457802] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe38aff6-be54-4522-ace8-d98c8caf6d22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.465030] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.465280] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.475680] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a376320a-52b3-4dfe-b212-08aae4fbff9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.479547] env[62066]: DEBUG nova.compute.provider_tree [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.488091] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 891.488091] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b01c35-f2d3-f421-690e-93e2efb1e8e5" [ 891.488091] env[62066]: _type = "Task" [ 891.488091] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.497335] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b01c35-f2d3-f421-690e-93e2efb1e8e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.529327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.553513] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.553773] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.704717] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ddabc59-f1db-423e-8808-7524e2969c96 req-7b674a3b-1c7b-4177-8a94-eb6fa2150977 service nova] Releasing lock "refresh_cache-b2afea5a-2390-40b0-b860-c50f71ebac94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.815533] env[62066]: DEBUG nova.objects.instance [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.942172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.983426] env[62066]: DEBUG nova.scheduler.client.report [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.997475] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b01c35-f2d3-f421-690e-93e2efb1e8e5, 'name': SearchDatastore_Task, 'duration_secs': 0.012858} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.998930] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2d22f01-283c-4a09-acec-daa3f00f47a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.005086] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 892.005086] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52169e99-594f-9436-0dd8-131203c98b19" [ 892.005086] env[62066]: _type = "Task" [ 892.005086] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.013100] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52169e99-594f-9436-0dd8-131203c98b19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.058599] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.063719] env[62066]: DEBUG nova.compute.manager [req-2bc16636-23cb-4da5-8e4b-6fc1837f7646 req-a36d52cf-08b1-4371-a50d-8cfa09a2396c service nova] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Received event network-vif-deleted-1ab1ec14-b82d-41c3-880e-3aed3469b3a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.321361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.321361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.321608] env[62066]: DEBUG nova.network.neutron [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.321608] env[62066]: DEBUG nova.objects.instance [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'info_cache' on Instance uuid ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.492803] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.493781] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.497860] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.423s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.500793] env[62066]: INFO nova.compute.claims [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.519341] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52169e99-594f-9436-0dd8-131203c98b19, 'name': SearchDatastore_Task, 'duration_secs': 0.031269} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.519930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.519930] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b2afea5a-2390-40b0-b860-c50f71ebac94/b2afea5a-2390-40b0-b860-c50f71ebac94.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.520214] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74c3de62-1c1f-44f4-bb07-d4040e0fc6bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.527190] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 892.527190] env[62066]: value = "task-1156364" [ 892.527190] env[62066]: _type = "Task" [ 892.527190] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.537065] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.584018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.825512] env[62066]: DEBUG nova.objects.base [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 893.006990] env[62066]: DEBUG nova.compute.utils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 893.010666] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.010666] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 893.037332] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156364, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.052859] env[62066]: DEBUG nova.policy [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.336992] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Successfully created port: b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.514021] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.553028] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156364, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.94519} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.554651] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] b2afea5a-2390-40b0-b860-c50f71ebac94/b2afea5a-2390-40b0-b860-c50f71ebac94.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.554919] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.555464] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b6beff9-a727-45d2-a253-bb7b72a0a6f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.564025] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 893.564025] env[62066]: value = "task-1156365" [ 893.564025] env[62066]: _type = "Task" [ 893.564025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.577505] env[62066]: DEBUG nova.network.neutron [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [{"id": "81a78519-507f-4de8-8e08-3a469f901c85", "address": "fa:16:3e:ab:3c:2d", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a78519-50", "ovs_interfaceid": "81a78519-507f-4de8-8e08-3a469f901c85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.578884] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.855360] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905386f0-1b87-48a4-9502-f0eda8ad49d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.864357] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b06d7b-7c6a-4fc0-89b3-be2befd10c6d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.900046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2060d80a-27cb-4604-9d7f-14ec45fc238c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.908731] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227436a0-3086-4145-9e5e-a5c18d11cecf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.926642] env[62066]: DEBUG nova.compute.provider_tree [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.076650] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090123} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.076944] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.077830] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e060a616-6597-45be-9517-b63e040ab2cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.080937] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ccb9f50f-dcc3-4d81-944e-d70803185ae1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.106792] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] b2afea5a-2390-40b0-b860-c50f71ebac94/b2afea5a-2390-40b0-b860-c50f71ebac94.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.107664] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e07dac7-0dc8-4edb-841f-b63c3b020536 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.130818] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 894.130818] env[62066]: value = "task-1156366" [ 894.130818] env[62066]: _type = "Task" [ 894.130818] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.140341] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.431943] env[62066]: DEBUG nova.scheduler.client.report [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.530557] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.555885] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.556388] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.556388] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.556523] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.556647] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.556800] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.557032] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.557238] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.557433] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.557606] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.557787] env[62066]: DEBUG nova.virt.hardware [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.558678] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69a33a2-eddd-4cf9-aa1e-d464fc58f8d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.567126] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c158fed1-4075-4e29-b16e-af18295ec41a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.584415] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.584693] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2753b919-899c-49f2-8f96-0d86d7c4226b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.591531] env[62066]: DEBUG oslo_vmware.api [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 894.591531] env[62066]: value = "task-1156367" [ 894.591531] env[62066]: _type = "Task" [ 894.591531] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.602118] env[62066]: DEBUG oslo_vmware.api [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.641046] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156366, 'name': ReconfigVM_Task, 'duration_secs': 0.308037} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.641248] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Reconfigured VM instance instance-00000055 to attach disk [datastore2] b2afea5a-2390-40b0-b860-c50f71ebac94/b2afea5a-2390-40b0-b860-c50f71ebac94.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.641919] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67fa0d34-ff81-4c29-9009-e4244a580535 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.648722] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 894.648722] env[62066]: value = "task-1156368" [ 894.648722] env[62066]: _type = "Task" [ 894.648722] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.656871] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156368, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.756689] env[62066]: DEBUG nova.compute.manager [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Received event network-vif-plugged-b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.756976] env[62066]: DEBUG oslo_concurrency.lockutils [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.757341] env[62066]: DEBUG oslo_concurrency.lockutils [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] Lock "7d7743d0-acf5-426e-b409-c00dca895048-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.757487] env[62066]: DEBUG oslo_concurrency.lockutils [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] Lock "7d7743d0-acf5-426e-b409-c00dca895048-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.757642] env[62066]: DEBUG nova.compute.manager [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] No waiting events found dispatching network-vif-plugged-b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.757788] env[62066]: WARNING nova.compute.manager [req-3e4164f8-84ea-495f-9412-0e3407b6ed1b req-ef221bbb-26ed-456d-ae0d-4753ad390aaf service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Received unexpected event network-vif-plugged-b010e41f-8864-4417-b170-3b94f1b1b74d for instance with vm_state building and task_state spawning. [ 894.855556] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Successfully updated port: b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.940357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.940946] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.944227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.737s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.944410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.944567] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 894.944896] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.829s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.949028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4931db62-27b7-47db-a58b-4503f51d4bfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.958050] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb77fe1-9c26-4698-ab53-eafa13e42b8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.978453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5499b9-d700-4ac4-90e4-3101e59e6dd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.993223] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89910205-7ac0-4263-a344-3e7315f8a317 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.030081] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179554MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 895.030802] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.104757] env[62066]: DEBUG oslo_vmware.api [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156367, 'name': PowerOnVM_Task, 'duration_secs': 0.421553} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.107212] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 895.107536] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.107777] env[62066]: DEBUG nova.compute.manager [None req-9f2d0de8-f51f-4943-a7da-5b775e6ce48e tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.108642] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc39d321-9171-46d5-9a9f-932430ec8022 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.113134] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6b4bf4-559a-416b-b685-8e8e629b7795 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.119740] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 895.119904] env[62066]: ERROR oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk due to incomplete transfer. [ 895.121562] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-02660df7-1560-44d0-b80e-baa58fa910de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.131255] env[62066]: DEBUG oslo_vmware.rw_handles [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c59d78-273b-d625-aaf2-1ab996920f25/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 895.131567] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Uploaded image c96c784e-9ccb-46d7-b666-d95d5fb00a8b to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 895.133937] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 895.134190] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b7af61c2-e7fc-43d8-9ad2-e862d2c90b5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.140886] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 895.140886] env[62066]: value = "task-1156369" [ 895.140886] env[62066]: _type = "Task" [ 895.140886] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.149096] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156369, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.161947] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156368, 'name': Rename_Task, 'duration_secs': 0.172095} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.162834] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.163103] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4652f3e-e4a7-49f7-978a-5b1fd3a9987b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.169573] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 895.169573] env[62066]: value = "task-1156370" [ 895.169573] env[62066]: _type = "Task" [ 895.169573] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.180184] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.361517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.361680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.361832] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.451542] env[62066]: DEBUG nova.compute.utils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 895.453069] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 895.453178] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.504711] env[62066]: DEBUG nova.policy [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91585bdab75c42178c835de912fd3fef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08737e043ede43b58298d1c1a834fa84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.654696] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156369, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.681883] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156370, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.726432] env[62066]: INFO nova.compute.manager [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Rebuilding instance [ 895.783055] env[62066]: DEBUG nova.compute.manager [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.783055] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8739a11-20a0-4626-adfb-008be646ef9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.789061] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaf5f8d-1468-4d37-be96-fe191edf7eaf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.800985] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192da737-1a18-474d-b312-5380410985f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.834997] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Successfully created port: f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 895.837522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f35c8b1-6fb7-487d-ad7a-7fecbeee00ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.846946] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca01d8e-b897-42d7-be01-954887ab9776 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.861675] env[62066]: DEBUG nova.compute.provider_tree [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.894096] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.956224] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.151954] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156369, 'name': Destroy_Task, 'duration_secs': 0.814543} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.152465] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Destroyed the VM [ 896.152737] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 896.153038] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3daeff3a-c3e0-4397-a0da-5772d47238ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.160817] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 896.160817] env[62066]: value = "task-1156371" [ 896.160817] env[62066]: _type = "Task" [ 896.160817] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.169521] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156371, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.179509] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156370, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.299200] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.299623] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd0d2532-fc7e-49e4-a5fb-f43a5379b478 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.307667] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 896.307667] env[62066]: value = "task-1156372" [ 896.307667] env[62066]: _type = "Task" [ 896.307667] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.320465] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.367046] env[62066]: DEBUG nova.scheduler.client.report [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.389492] env[62066]: DEBUG nova.network.neutron [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Updating instance_info_cache with network_info: [{"id": "b010e41f-8864-4417-b170-3b94f1b1b74d", "address": "fa:16:3e:db:67:2c", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb010e41f-88", "ovs_interfaceid": "b010e41f-8864-4417-b170-3b94f1b1b74d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.672303] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156371, 'name': RemoveSnapshot_Task, 'duration_secs': 0.383833} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.675530] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 896.675910] env[62066]: DEBUG nova.compute.manager [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.676628] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc51982-bc76-47b7-b83e-13dfc43d1d8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.683894] env[62066]: DEBUG oslo_vmware.api [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156370, 'name': PowerOnVM_Task, 'duration_secs': 1.148224} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.685463] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.685674] env[62066]: INFO nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Took 9.16 seconds to spawn the instance on the hypervisor. [ 896.685855] env[62066]: DEBUG nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.688680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d47a82-b3c1-4567-b8fd-00b77095b747 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.821497] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156372, 'name': PowerOffVM_Task, 'duration_secs': 0.226451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.821497] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.821747] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.821814] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4d8881a-f7e9-4645-9ce9-73ecd0e64c82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.828028] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 896.828028] env[62066]: value = "task-1156373" [ 896.828028] env[62066]: _type = "Task" [ 896.828028] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.835602] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156373, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.866570] env[62066]: DEBUG nova.compute.manager [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Received event network-changed-b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.866756] env[62066]: DEBUG nova.compute.manager [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Refreshing instance network info cache due to event network-changed-b010e41f-8864-4417-b170-3b94f1b1b74d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.866950] env[62066]: DEBUG oslo_concurrency.lockutils [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] Acquiring lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.892419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.892719] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Instance network_info: |[{"id": "b010e41f-8864-4417-b170-3b94f1b1b74d", "address": "fa:16:3e:db:67:2c", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb010e41f-88", "ovs_interfaceid": "b010e41f-8864-4417-b170-3b94f1b1b74d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.893030] env[62066]: DEBUG oslo_concurrency.lockutils [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] Acquired lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.893268] env[62066]: DEBUG nova.network.neutron [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Refreshing network info cache for port b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.896750] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:67:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b010e41f-8864-4417-b170-3b94f1b1b74d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.902633] env[62066]: DEBUG oslo.service.loopingcall [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.905327] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.905781] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d433d405-a326-4d44-9820-026241924156 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.926097] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.926097] env[62066]: value = "task-1156374" [ 896.926097] env[62066]: _type = "Task" [ 896.926097] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.937822] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156374, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.967654] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.992756] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.993014] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.993181] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.993368] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.993513] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.993660] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.993863] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.994034] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.994270] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.994466] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.994639] env[62066]: DEBUG nova.virt.hardware [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.995526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793fafd0-b770-41e5-b1e0-eb7c247cc043 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.005430] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b49216-a799-4a87-8e22-fa6587ddee0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.118669] env[62066]: DEBUG nova.network.neutron [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Updated VIF entry in instance network info cache for port b010e41f-8864-4417-b170-3b94f1b1b74d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.119092] env[62066]: DEBUG nova.network.neutron [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Updating instance_info_cache with network_info: [{"id": "b010e41f-8864-4417-b170-3b94f1b1b74d", "address": "fa:16:3e:db:67:2c", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb010e41f-88", "ovs_interfaceid": "b010e41f-8864-4417-b170-3b94f1b1b74d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.197238] env[62066]: INFO nova.compute.manager [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Shelve offloading [ 897.203663] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.203663] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f2c59a4-17a3-4a27-a6d5-fa75419091a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.208034] env[62066]: INFO nova.compute.manager [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Took 14.28 seconds to build instance. [ 897.212152] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 897.212152] env[62066]: value = "task-1156375" [ 897.212152] env[62066]: _type = "Task" [ 897.212152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.221178] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 897.221407] env[62066]: DEBUG nova.compute.manager [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.222213] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0701d8-0861-4f2a-a5a7-589bf5bdb52b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.225199] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.225509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.225648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.225858] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.226056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.228408] env[62066]: INFO nova.compute.manager [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Terminating instance [ 897.231593] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.231774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.231958] env[62066]: DEBUG nova.network.neutron [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.233546] env[62066]: DEBUG nova.compute.manager [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.233630] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.234437] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b494399-5065-498f-be00-a8d5b05d9088 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.242290] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.242611] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75c4fd56-6bdb-4b87-b249-4f56c55b70d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.250062] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 897.250062] env[62066]: value = "task-1156376" [ 897.250062] env[62066]: _type = "Task" [ 897.250062] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.258492] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.326163] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Successfully updated port: f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.339380] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 897.339648] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 897.339878] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251692', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'name': 'volume-1f4e0b34-91c4-413f-8225-3151a6e125ab', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753f585a-22d7-4eeb-8580-4e3a68b5fd72', 'attached_at': '', 'detached_at': '', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'serial': '1f4e0b34-91c4-413f-8225-3151a6e125ab'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 897.341026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1d316a-7fe0-4191-91ce-3974aef94dbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.361147] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee80fc5-c7b3-4ffa-a64b-62c0ff4305b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.369106] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce93c73-c15b-426a-8ab3-6c0b59e9f5be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.388432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.443s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.392104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.571s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.393333] env[62066]: INFO nova.compute.claims [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.397034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449ca7e7-bfb7-444e-b04b-1131dd50630a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.412728] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] The volume has not been displaced from its original location: [datastore2] volume-1f4e0b34-91c4-413f-8225-3151a6e125ab/volume-1f4e0b34-91c4-413f-8225-3151a6e125ab.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 897.418054] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Reconfiguring VM instance instance-0000004c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 897.418433] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-242aa862-554e-4a5e-81e9-89cc1b5b827f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.444824] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 897.444824] env[62066]: value = "task-1156377" [ 897.444824] env[62066]: _type = "Task" [ 897.444824] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.445180] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156374, 'name': CreateVM_Task, 'duration_secs': 0.379508} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.445459] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.448949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.449126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.449439] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.449716] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dfafbe1-e48b-4a09-b687-9580cb843487 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.454756] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 897.454756] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52eda96a-38ec-6a8f-b30d-d9225bfd16e0" [ 897.454756] env[62066]: _type = "Task" [ 897.454756] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.458206] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156377, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.468542] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52eda96a-38ec-6a8f-b30d-d9225bfd16e0, 'name': SearchDatastore_Task, 'duration_secs': 0.01046} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.468815] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.469121] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.469381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.469531] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.469711] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.469962] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c528149-d276-40fa-b92e-a9730147f3f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.477679] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.477867] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.478584] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfdeb05c-2c9a-4923-8544-2abb9496ae4f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.483377] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 897.483377] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522ae05b-d4ca-729a-4e60-6fb690a85ce9" [ 897.483377] env[62066]: _type = "Task" [ 897.483377] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.490665] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522ae05b-d4ca-729a-4e60-6fb690a85ce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.622427] env[62066]: DEBUG oslo_concurrency.lockutils [req-fd851ec4-5c6c-4066-92b7-94aefb0a720c req-0f113595-cb0e-49a9-98dc-3eef9d329526 service nova] Releasing lock "refresh_cache-7d7743d0-acf5-426e-b409-c00dca895048" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.711296] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08aa8042-b8e4-42f0-a6e0-b9b76e321b10 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.794s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.760076] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156376, 'name': PowerOffVM_Task, 'duration_secs': 0.192558} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.760436] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.760622] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.760881] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1490007-4d50-4bda-9b47-259adf175f52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.824818] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.825081] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.825307] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleting the datastore file [datastore1] ccb9f50f-dcc3-4d81-944e-d70803185ae1 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.825581] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-735f98be-531a-41cf-b6dc-185ae854b61a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.828032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.828163] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.828343] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.834692] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 897.834692] env[62066]: value = "task-1156379" [ 897.834692] env[62066]: _type = "Task" [ 897.834692] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.845492] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156379, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.946139] env[62066]: DEBUG nova.network.neutron [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.956614] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156377, 'name': ReconfigVM_Task, 'duration_secs': 0.434596} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.957528] env[62066]: INFO nova.scheduler.client.report [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocation for migration 45cece1f-bf1a-4a05-81c8-de8c7a858bdd [ 897.958419] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Reconfigured VM instance instance-0000004c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 897.965510] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a29b8b79-fada-4904-8467-9735718d1d8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.981794] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 897.981794] env[62066]: value = "task-1156380" [ 897.981794] env[62066]: _type = "Task" [ 897.981794] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.996084] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156380, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.999789] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522ae05b-d4ca-729a-4e60-6fb690a85ce9, 'name': SearchDatastore_Task, 'duration_secs': 0.007921} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.001295] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50d5aef2-1862-47df-b0bc-8831cbe0b196 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.011726] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 898.011726] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529dfc23-1464-769a-e5ca-bb09f6a9ccaa" [ 898.011726] env[62066]: _type = "Task" [ 898.011726] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.021196] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529dfc23-1464-769a-e5ca-bb09f6a9ccaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.344894] env[62066]: DEBUG oslo_vmware.api [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156379, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141874} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.345213] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.345462] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.345712] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.345869] env[62066]: INFO nova.compute.manager [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 898.346127] env[62066]: DEBUG oslo.service.loopingcall [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.346328] env[62066]: DEBUG nova.compute.manager [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.346463] env[62066]: DEBUG nova.network.neutron [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 898.368397] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 898.451778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.477716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ff1eba2-b4c7-4d70-9b1d-bda4da4f3d21 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.579s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.505815] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156380, 'name': ReconfigVM_Task, 'duration_secs': 0.242784} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.506142] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251692', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'name': 'volume-1f4e0b34-91c4-413f-8225-3151a6e125ab', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '753f585a-22d7-4eeb-8580-4e3a68b5fd72', 'attached_at': '', 'detached_at': '', 'volume_id': '1f4e0b34-91c4-413f-8225-3151a6e125ab', 'serial': '1f4e0b34-91c4-413f-8225-3151a6e125ab'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 898.506633] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.507268] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb461ef9-d64b-4e89-957e-1f91e7e80a76 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.516101] env[62066]: DEBUG nova.network.neutron [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating instance_info_cache with network_info: [{"id": "f0fae426-1b77-435b-98a6-72d487ff589f", "address": "fa:16:3e:8c:82:6e", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fae426-1b", "ovs_interfaceid": "f0fae426-1b77-435b-98a6-72d487ff589f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.524820] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.528387] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d3b804b-d15d-438c-a56d-a5f59733dcef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.532498] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529dfc23-1464-769a-e5ca-bb09f6a9ccaa, 'name': SearchDatastore_Task, 'duration_secs': 0.008947} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.535731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.536050] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 7d7743d0-acf5-426e-b409-c00dca895048/7d7743d0-acf5-426e-b409-c00dca895048.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.536917] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de055e32-85a9-4f9f-b3b2-b6250f260915 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.544492] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 898.544492] env[62066]: value = "task-1156382" [ 898.544492] env[62066]: _type = "Task" [ 898.544492] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.556239] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.597789] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.598184] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.598409] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Deleting the datastore file [datastore2] 753f585a-22d7-4eeb-8580-4e3a68b5fd72 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.598684] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-292b9930-86eb-4375-828c-77dd82898e70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.608517] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for the task: (returnval){ [ 898.608517] env[62066]: value = "task-1156383" [ 898.608517] env[62066]: _type = "Task" [ 898.608517] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.616121] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.698188] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8bdcf11-ec72-4841-aaa0-50df81caf47c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.706762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62f1e57-9d47-4f33-a1fa-03df0d39db4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.741808] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a5bfe6-d441-4ee8-b3fc-7a63efa98d45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.750942] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3dbca36-06ed-498e-8794-26e417230e2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.767860] env[62066]: DEBUG nova.compute.provider_tree [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.972659] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "4983ece5-bb4b-43cb-a928-324b2f31030e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.973141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.022277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.022644] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance network_info: |[{"id": "f0fae426-1b77-435b-98a6-72d487ff589f", "address": "fa:16:3e:8c:82:6e", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fae426-1b", "ovs_interfaceid": "f0fae426-1b77-435b-98a6-72d487ff589f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.023189] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:82:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0fae426-1b77-435b-98a6-72d487ff589f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.032717] env[62066]: DEBUG oslo.service.loopingcall [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.034053] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.034053] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-872bc897-d5ab-4d43-b586-105506126caa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.063297] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156382, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.064512] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.064512] env[62066]: value = "task-1156384" [ 899.064512] env[62066]: _type = "Task" [ 899.064512] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.072778] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156384, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.118587] env[62066]: DEBUG oslo_vmware.api [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Task: {'id': task-1156383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080326} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.118855] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.119072] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.119270] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.184143] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 899.184143] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f71af56-02c1-4443-817f-016c9ed12b18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.194963] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea843b6-cbfe-4967-bb7c-a82a2b925f3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.213153] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.214252] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ec5a6f-464f-4689-ac30-8e60eb63ab46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.225638] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.226210] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-933efe7b-b691-4b11-970e-060e820d2b5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.244477] env[62066]: ERROR nova.compute.manager [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Failed to detach volume 1f4e0b34-91c4-413f-8225-3151a6e125ab from /dev/sda: nova.exception.InstanceNotFound: Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 could not be found. [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Traceback (most recent call last): [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self.driver.rebuild(**kwargs) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise NotImplementedError() [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] NotImplementedError [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] During handling of the above exception, another exception occurred: [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Traceback (most recent call last): [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self.driver.detach_volume(context, old_connection_info, [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] return self._volumeops.detach_volume(connection_info, instance) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._detach_volume_vmdk(connection_info, instance) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] stable_ref.fetch_moref(session) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise exception.InstanceNotFound(instance_id=self._uuid) [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] nova.exception.InstanceNotFound: Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 could not be found. [ 899.244477] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.250562] env[62066]: DEBUG nova.compute.manager [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Received event network-vif-plugged-f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.250824] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.255056] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.255359] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.004s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.255512] env[62066]: DEBUG nova.compute.manager [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] No waiting events found dispatching network-vif-plugged-f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.255688] env[62066]: WARNING nova.compute.manager [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Received unexpected event network-vif-plugged-f0fae426-1b77-435b-98a6-72d487ff589f for instance with vm_state building and task_state spawning. [ 899.255851] env[62066]: DEBUG nova.compute.manager [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Received event network-changed-f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.256013] env[62066]: DEBUG nova.compute.manager [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Refreshing instance network info cache due to event network-changed-f0fae426-1b77-435b-98a6-72d487ff589f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.256224] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Acquiring lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.256389] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Acquired lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.256597] env[62066]: DEBUG nova.network.neutron [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Refreshing network info cache for port f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.272811] env[62066]: DEBUG nova.scheduler.client.report [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.327497] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.327753] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.327924] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleting the datastore file [datastore2] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.328240] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d22a42a0-c8c0-4dd6-8085-94c8a9f0f3fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.340039] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 899.340039] env[62066]: value = "task-1156386" [ 899.340039] env[62066]: _type = "Task" [ 899.340039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.351514] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.361542] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.361768] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.361957] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.362247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.362472] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.368634] env[62066]: INFO nova.compute.manager [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Terminating instance [ 899.370716] env[62066]: DEBUG nova.compute.manager [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.370954] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.371940] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffadd6fb-f396-47c0-8edb-80a46a562e93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.385333] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.385661] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-283b16b6-3819-4f71-809d-3154180c8f57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.394280] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 899.394280] env[62066]: value = "task-1156387" [ 899.394280] env[62066]: _type = "Task" [ 899.394280] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.405284] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.455788] env[62066]: DEBUG nova.compute.utils [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Build of instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 aborted: Failed to rebuild volume backed instance. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 899.458585] env[62066]: ERROR nova.compute.manager [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 aborted: Failed to rebuild volume backed instance. [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Traceback (most recent call last): [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self.driver.rebuild(**kwargs) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise NotImplementedError() [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] NotImplementedError [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] During handling of the above exception, another exception occurred: [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Traceback (most recent call last): [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._detach_root_volume(context, instance, root_bdm) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] with excutils.save_and_reraise_exception(): [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self.force_reraise() [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise self.value [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self.driver.detach_volume(context, old_connection_info, [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] return self._volumeops.detach_volume(connection_info, instance) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._detach_volume_vmdk(connection_info, instance) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] stable_ref.fetch_moref(session) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise exception.InstanceNotFound(instance_id=self._uuid) [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] nova.exception.InstanceNotFound: Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 could not be found. [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] During handling of the above exception, another exception occurred: [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Traceback (most recent call last): [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] yield [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 899.458585] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._do_rebuild_instance_with_claim( [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._do_rebuild_instance( [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._rebuild_default_impl(**kwargs) [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] self._rebuild_volume_backed_instance( [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] raise exception.BuildAbortException( [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] nova.exception.BuildAbortException: Build of instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 aborted: Failed to rebuild volume backed instance. [ 899.460719] env[62066]: ERROR nova.compute.manager [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] [ 899.476471] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.507679] env[62066]: DEBUG nova.compute.manager [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-vif-unplugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.507963] env[62066]: DEBUG oslo_concurrency.lockutils [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.508321] env[62066]: DEBUG oslo_concurrency.lockutils [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.508375] env[62066]: DEBUG oslo_concurrency.lockutils [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.508563] env[62066]: DEBUG nova.compute.manager [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] No waiting events found dispatching network-vif-unplugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.508724] env[62066]: WARNING nova.compute.manager [req-e722e653-8cbd-4806-9924-886051b05a73 req-0d6c3aa8-2ca1-4312-92ac-6c16b9a7b808 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received unexpected event network-vif-unplugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef for instance with vm_state shelved and task_state shelving_offloading. [ 899.563793] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.563990] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 7d7743d0-acf5-426e-b409-c00dca895048/7d7743d0-acf5-426e-b409-c00dca895048.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.564219] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.564600] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75a61f0f-a585-4c3b-93d7-547d5ff15148 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.578833] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156384, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.580589] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 899.580589] env[62066]: value = "task-1156388" [ 899.580589] env[62066]: _type = "Task" [ 899.580589] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.593417] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156388, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.778097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.778789] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.781208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.252s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.781463] env[62066]: DEBUG nova.objects.instance [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lazy-loading 'resources' on Instance uuid 0fda5f59-55ac-4150-8402-00064d14c8ab {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.852501] env[62066]: DEBUG oslo_vmware.api [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183073} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.853097] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.853097] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.853191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.857852] env[62066]: DEBUG nova.network.neutron [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.884275] env[62066]: INFO nova.scheduler.client.report [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted allocations for instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 [ 899.909243] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156387, 'name': PowerOffVM_Task, 'duration_secs': 0.246243} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.909575] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.909957] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.910290] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fa3dca3-bfc6-46ff-9b6d-393e2336ede9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.985641] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.985820] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.986015] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleting the datastore file [datastore2] a2251e56-2787-412d-89c9-eef111ee6d2b {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.988504] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d230b98e-ff32-4ed5-9d74-665f63bd4a8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.998246] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for the task: (returnval){ [ 899.998246] env[62066]: value = "task-1156390" [ 899.998246] env[62066]: _type = "Task" [ 899.998246] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.003189] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.009235] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.077241] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156384, 'name': CreateVM_Task, 'duration_secs': 0.584165} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.077468] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.078174] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.078350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.078680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.078953] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-418b9deb-835d-4c7c-b545-2d2d97dae08c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.085655] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 900.085655] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52899e1c-3b0a-ef27-591f-0405b0c7bf65" [ 900.085655] env[62066]: _type = "Task" [ 900.085655] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.092177] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156388, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074149} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.092795] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.093666] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f185b9f-3695-4666-b02c-de7849c48471 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.099545] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52899e1c-3b0a-ef27-591f-0405b0c7bf65, 'name': SearchDatastore_Task, 'duration_secs': 0.011493} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.100168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.100412] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.100685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.100847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.101164] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.101405] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47909e06-4eb7-404d-b836-812383e6289a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.122814] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 7d7743d0-acf5-426e-b409-c00dca895048/7d7743d0-acf5-426e-b409-c00dca895048.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.126330] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b70eb101-80f0-4dc4-ad04-a40e9b6a2554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.142492] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.142688] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.143771] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5ec5f16-6eda-4641-b774-56f8c8122ab5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.149779] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 900.149779] env[62066]: value = "task-1156391" [ 900.149779] env[62066]: _type = "Task" [ 900.149779] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.151132] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 900.151132] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5251aa87-2919-51c6-c42d-810440f09742" [ 900.151132] env[62066]: _type = "Task" [ 900.151132] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.162435] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5251aa87-2919-51c6-c42d-810440f09742, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.165096] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156391, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.228356] env[62066]: DEBUG nova.network.neutron [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updated VIF entry in instance network info cache for port f0fae426-1b77-435b-98a6-72d487ff589f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.228779] env[62066]: DEBUG nova.network.neutron [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating instance_info_cache with network_info: [{"id": "f0fae426-1b77-435b-98a6-72d487ff589f", "address": "fa:16:3e:8c:82:6e", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fae426-1b", "ovs_interfaceid": "f0fae426-1b77-435b-98a6-72d487ff589f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.284435] env[62066]: DEBUG nova.compute.utils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 900.285981] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.286209] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 900.363518] env[62066]: INFO nova.compute.manager [-] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Took 2.02 seconds to deallocate network for instance. [ 900.371980] env[62066]: DEBUG nova.policy [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd53366fea3a6434fadadede81df16089', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '33b756ddd30f4cb0b917fad171213266', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.387985] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.512979] env[62066]: DEBUG oslo_vmware.api [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Task: {'id': task-1156390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146206} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.514954] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.515191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 900.515368] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 900.515539] env[62066]: INFO nova.compute.manager [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 900.515776] env[62066]: DEBUG oslo.service.loopingcall [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.518277] env[62066]: DEBUG nova.compute.manager [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.518382] env[62066]: DEBUG nova.network.neutron [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.604295] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfc7199-3b52-4d03-abe1-c79b4a47f40c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.614652] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30c52c4-376b-4b67-8374-95c8ba054c37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.648772] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c001ba88-accd-487c-b285-9c7480327ac3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.664247] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2695e655-cb18-421a-8063-c685a8f71286 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.668012] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156391, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.671629] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5251aa87-2919-51c6-c42d-810440f09742, 'name': SearchDatastore_Task, 'duration_secs': 0.010932} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.672755] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fb91f9f-dae7-439b-8f48-d1d2579ee8c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.683144] env[62066]: DEBUG nova.compute.provider_tree [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.690831] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 900.690831] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522a3c26-a566-451b-8c36-658ef0d50d3e" [ 900.690831] env[62066]: _type = "Task" [ 900.690831] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.698132] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522a3c26-a566-451b-8c36-658ef0d50d3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.737308] env[62066]: DEBUG oslo_concurrency.lockutils [req-0ac7feaf-cb3f-4f30-83f1-973f07a84db8 req-fddeee74-5e15-40ab-a385-633fae8611f3 service nova] Releasing lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.793494] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.870922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.166909] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156391, 'name': ReconfigVM_Task, 'duration_secs': 0.752691} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.167271] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 7d7743d0-acf5-426e-b409-c00dca895048/7d7743d0-acf5-426e-b409-c00dca895048.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.167974] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86ed67bd-0190-4195-80ff-e8c777c72408 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.175323] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 901.175323] env[62066]: value = "task-1156392" [ 901.175323] env[62066]: _type = "Task" [ 901.175323] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.183482] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156392, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.186430] env[62066]: DEBUG nova.scheduler.client.report [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.198886] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522a3c26-a566-451b-8c36-658ef0d50d3e, 'name': SearchDatastore_Task, 'duration_secs': 0.010817} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.199435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.200296] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.200296] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-409902a7-8472-4031-a991-94c121b00b9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.208442] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 901.208442] env[62066]: value = "task-1156393" [ 901.208442] env[62066]: _type = "Task" [ 901.208442] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.217806] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.474417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.688455] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156392, 'name': Rename_Task, 'duration_secs': 0.168261} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.688818] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.689143] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea49a45c-0aa1-4186-a334-6881fc7222ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.698778] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 901.698778] env[62066]: value = "task-1156394" [ 901.698778] env[62066]: _type = "Task" [ 901.698778] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.703429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.705699] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.764s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.705990] env[62066]: DEBUG nova.objects.instance [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lazy-loading 'resources' on Instance uuid 854a1ea4-515d-4a05-b179-22713b63f7c3 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.721673] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.729837] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489192} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.730172] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.730436] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.730714] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0edb392-4abe-43b2-88da-96b417f8a892 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.739206] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 901.739206] env[62066]: value = "task-1156395" [ 901.739206] env[62066]: _type = "Task" [ 901.739206] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.746641] env[62066]: INFO nova.scheduler.client.report [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted allocations for instance 0fda5f59-55ac-4150-8402-00064d14c8ab [ 901.751620] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156395, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.752615] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Successfully created port: d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.805636] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.832161] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.832773] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.833134] env[62066]: DEBUG nova.virt.hardware [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.834547] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66168854-519d-43ea-9fac-0c5e6132bd0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.843285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8146ed3d-e616-48a1-895b-c51ac25b873f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.214218] env[62066]: DEBUG oslo_vmware.api [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156394, 'name': PowerOnVM_Task, 'duration_secs': 0.460517} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.215249] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.215778] env[62066]: INFO nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Took 7.69 seconds to spawn the instance on the hypervisor. [ 902.217017] env[62066]: DEBUG nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.217361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e58f73-1cf8-4656-96bd-e7af43562d50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.233447] env[62066]: DEBUG nova.compute.manager [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.233804] env[62066]: DEBUG nova.compute.manager [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing instance network info cache due to event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.234435] env[62066]: DEBUG oslo_concurrency.lockutils [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.234728] env[62066]: DEBUG oslo_concurrency.lockutils [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.235059] env[62066]: DEBUG nova.network.neutron [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.262136] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156395, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072729} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.262136] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.262136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf9dd07-861e-4ca3-be12-40cb414d8b5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.268615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c5c0db8b-54b8-4f59-972b-f6bcc67e0fbf tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "0fda5f59-55ac-4150-8402-00064d14c8ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.762s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.270262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.303405] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.307903] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e674e6a3-5ba4-4e95-a59e-2065a3597d9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.330749] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 902.330749] env[62066]: value = "task-1156396" [ 902.330749] env[62066]: _type = "Task" [ 902.330749] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.346787] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.590568] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4a64cd-fe7e-4bd2-aaf0-9eac5f4ee5ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.600463] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05d6669-f98b-45f2-923b-20fa2760899d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.635582] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a2f5b3-a0c7-47ff-91a6-d4b70b8a62f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.645732] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c594e9-df7e-43ac-a418-1b8cd6caebcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.661396] env[62066]: DEBUG nova.compute.provider_tree [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.749654] env[62066]: INFO nova.compute.manager [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Took 18.57 seconds to build instance. [ 902.783843] env[62066]: DEBUG nova.network.neutron [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.841065] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.971770] env[62066]: DEBUG nova.network.neutron [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updated VIF entry in instance network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.972187] env[62066]: DEBUG nova.network.neutron [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.168071] env[62066]: DEBUG nova.scheduler.client.report [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.171687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.171924] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.172276] env[62066]: DEBUG nova.objects.instance [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.251725] env[62066]: DEBUG oslo_concurrency.lockutils [None req-07deeca5-483e-4ba0-b42c-0cc491fba5af tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.082s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.286938] env[62066]: INFO nova.compute.manager [-] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Took 2.77 seconds to deallocate network for instance. [ 903.343242] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156396, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.377574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.377851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.378109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.378381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.378637] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.381839] env[62066]: INFO nova.compute.manager [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Terminating instance [ 903.387055] env[62066]: DEBUG nova.compute.manager [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.387277] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.388154] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c097a66-0068-44ac-9514-e15f27ed728e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.397124] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.397386] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f34e1d0-62ea-4e89-bbed-473d86d35ad9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.403954] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 903.403954] env[62066]: value = "task-1156397" [ 903.403954] env[62066]: _type = "Task" [ 903.403954] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.411420] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.463543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.463883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.464214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.464457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.464642] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.467177] env[62066]: INFO nova.compute.manager [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Terminating instance [ 903.471188] env[62066]: DEBUG nova.compute.manager [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.471310] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a260fb60-acab-4bfc-b00e-e56b128be2e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.475367] env[62066]: DEBUG oslo_concurrency.lockutils [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.475555] env[62066]: DEBUG nova.compute.manager [req-5c43e7ff-2e48-4500-bd26-d8a60d4e2497 req-23b8f536-709c-46bc-966b-04a90fd4e1c5 service nova] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Received event network-vif-deleted-81a78519-507f-4de8-8e08-3a469f901c85 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.481337] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093cab70-e254-48f0-8d5c-f4f791a27bd2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.519084] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 could not be found. [ 903.519391] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.519740] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92453b80-17c6-472e-b97c-92f8218a5e6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.528771] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0a64f4-5433-4b52-b61c-eff11c293be9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.559433] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 753f585a-22d7-4eeb-8580-4e3a68b5fd72 could not be found. [ 903.559649] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.559832] env[62066]: INFO nova.compute.manager [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Took 0.09 seconds to destroy the instance on the hypervisor. [ 903.560108] env[62066]: DEBUG oslo.service.loopingcall [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.560372] env[62066]: DEBUG nova.compute.manager [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.560486] env[62066]: DEBUG nova.network.neutron [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.675093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.679960] env[62066]: DEBUG nova.objects.instance [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.682038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.098s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.682038] env[62066]: INFO nova.compute.claims [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.701778] env[62066]: INFO nova.scheduler.client.report [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted allocations for instance 854a1ea4-515d-4a05-b179-22713b63f7c3 [ 903.793740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.842205] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156396, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.913166] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156397, 'name': PowerOffVM_Task, 'duration_secs': 0.196006} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.913534] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.913711] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.913979] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f676a20-cbad-462d-91da-e81effaa756f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.979082] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.979490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.979791] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore1] ecf9fd16-82c7-4bea-b6a9-7262e75effef {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.980284] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f3da03d-05b6-4dc1-abd5-5e43290c2cda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.988869] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 903.988869] env[62066]: value = "task-1156399" [ 903.988869] env[62066]: _type = "Task" [ 903.988869] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.998609] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156399, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.055728] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.055993] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.099038] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Successfully updated port: d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.185982] env[62066]: DEBUG nova.objects.base [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.186270] env[62066]: DEBUG nova.network.neutron [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.212019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2b0b0e27-ea26-48c5-b703-d830a4384662 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "854a1ea4-515d-4a05-b179-22713b63f7c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.264s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.344724] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156396, 'name': ReconfigVM_Task, 'duration_secs': 1.885476} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.345017] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.345650] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb96d9df-ac76-401e-a8aa-c3020e0c976c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.352298] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 904.352298] env[62066]: value = "task-1156400" [ 904.352298] env[62066]: _type = "Task" [ 904.352298] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.360210] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156400, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.367505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8caebb6c-157a-48a3-8dc7-750fc87f58b1 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.196s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.453941] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.454225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.500441] env[62066]: DEBUG oslo_vmware.api [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156399, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.500629] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.500819] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.501015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.501210] env[62066]: INFO nova.compute.manager [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Took 1.11 seconds to destroy the instance on the hypervisor. [ 904.501474] env[62066]: DEBUG oslo.service.loopingcall [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.501671] env[62066]: DEBUG nova.compute.manager [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.501766] env[62066]: DEBUG nova.network.neutron [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 904.558766] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.601888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.602068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.602298] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.862851] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156400, 'name': Rename_Task, 'duration_secs': 0.227781} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.865739] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.866194] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87437570-d793-42bd-a9f4-9aa96f767864 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.872783] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 904.872783] env[62066]: value = "task-1156401" [ 904.872783] env[62066]: _type = "Task" [ 904.872783] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.882262] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.915667] env[62066]: DEBUG nova.network.neutron [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.957081] env[62066]: DEBUG nova.compute.utils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 904.975353] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a66b82b-9c30-4d09-b27d-6971cb11e6ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.983849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a37328-98a9-48b7-ba29-653c8d1835b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.015334] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61125d0d-40c3-419d-a100-e0b0ae8428be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.022900] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147017fd-476c-486f-8261-bec927c7447a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.038968] env[62066]: DEBUG nova.compute.provider_tree [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.078317] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.135657] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.242156] env[62066]: DEBUG nova.network.neutron [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.274181] env[62066]: DEBUG nova.network.neutron [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Updating instance_info_cache with network_info: [{"id": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "address": "fa:16:3e:3c:0e:e6", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1a44472-08", "ovs_interfaceid": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.384923] env[62066]: DEBUG oslo_vmware.api [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156401, 'name': PowerOnVM_Task, 'duration_secs': 0.466979} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.385197] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.385507] env[62066]: INFO nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Took 8.42 seconds to spawn the instance on the hypervisor. [ 905.385584] env[62066]: DEBUG nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.386343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c8a102-1047-4827-ab5b-f8f102916912 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.417738] env[62066]: INFO nova.compute.manager [-] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Took 1.86 seconds to deallocate network for instance. [ 905.460162] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.542067] env[62066]: DEBUG nova.scheduler.client.report [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.724110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.724368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.745420] env[62066]: INFO nova.compute.manager [-] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Took 1.24 seconds to deallocate network for instance. [ 905.777299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.777607] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Instance network_info: |[{"id": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "address": "fa:16:3e:3c:0e:e6", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1a44472-08", "ovs_interfaceid": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.778014] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:0e:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1a44472-08ab-45ac-94eb-3a8b9bb7ee35', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.786033] env[62066]: DEBUG oslo.service.loopingcall [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.786487] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.786727] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17b75e59-7bfa-45b0-93fe-f5222b4cba0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.807382] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.807382] env[62066]: value = "task-1156402" [ 905.807382] env[62066]: _type = "Task" [ 905.807382] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.815243] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156402, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.902020] env[62066]: INFO nova.compute.manager [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Took 19.85 seconds to build instance. [ 905.963740] env[62066]: INFO nova.compute.manager [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Took 0.55 seconds to detach 1 volumes for instance. [ 905.966050] env[62066]: DEBUG nova.compute.manager [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Deleting volume: 1f4e0b34-91c4-413f-8225-3151a6e125ab {{(pid=62066) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 906.047298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.048166] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.053559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.023s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.178382] env[62066]: DEBUG nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Received event network-vif-deleted-4991ab48-c8b7-437b-81d7-db93c7d3da4d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.178382] env[62066]: DEBUG nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Received event network-vif-plugged-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.178534] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Acquiring lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.178758] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.178918] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.179099] env[62066]: DEBUG nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] No waiting events found dispatching network-vif-plugged-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.179274] env[62066]: WARNING nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Received unexpected event network-vif-plugged-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 for instance with vm_state building and task_state spawning. [ 906.179474] env[62066]: DEBUG nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Received event network-changed-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.179594] env[62066]: DEBUG nova.compute.manager [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Refreshing instance network info cache due to event network-changed-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.179789] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Acquiring lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.179927] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Acquired lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.180104] env[62066]: DEBUG nova.network.neutron [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Refreshing network info cache for port d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.228464] env[62066]: DEBUG nova.compute.utils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 906.251180] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.317525] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156402, 'name': CreateVM_Task, 'duration_secs': 0.453061} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.317718] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.318484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.318667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.319006] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.319280] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd9b6706-27f4-4f32-85b1-344d09371ebd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.324173] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 906.324173] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528b9e75-7eb7-295e-0400-7eb0703cc096" [ 906.324173] env[62066]: _type = "Task" [ 906.324173] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.332259] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528b9e75-7eb7-295e-0400-7eb0703cc096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.404800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54864319-ea7d-4835-b229-105bfac275d1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.358s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.510941] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.528902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.529206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.529454] env[62066]: INFO nova.compute.manager [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Attaching volume cd885df2-f1e2-4754-bb18-08a36f7f908d to /dev/sdb [ 906.558731] env[62066]: DEBUG nova.compute.utils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 906.568907] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 906.569097] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 906.571804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd1652b-4d21-4e01-9525-ace23344c462 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.580514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ab9bb1-626c-41d6-af86-f8bd81d33c50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.594047] env[62066]: DEBUG nova.virt.block_device [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Updating existing volume attachment record: 00b5334c-d529-4fe2-a6cc-cf403200abb5 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 906.632503] env[62066]: DEBUG nova.policy [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b74a303468b4af3ac650bab8d810d10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcef20dbc27940be90f355cc136bd436', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.736716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.836981] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528b9e75-7eb7-295e-0400-7eb0703cc096, 'name': SearchDatastore_Task, 'duration_secs': 0.015454} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.837337] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.837577] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.837811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.837958] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.838154] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.838935] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e070d1ad-e857-4de1-ab70-e09aceb802a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.847064] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.847432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.848147] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39c3a3aa-af5a-44ca-a80e-70d7ef62a686 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.853484] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 906.853484] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528fd9a5-97c3-bfbd-ef07-cc41724bd4ce" [ 906.853484] env[62066]: _type = "Task" [ 906.853484] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.861441] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528fd9a5-97c3-bfbd-ef07-cc41724bd4ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.042361] env[62066]: DEBUG nova.network.neutron [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Updated VIF entry in instance network info cache for port d1a44472-08ab-45ac-94eb-3a8b9bb7ee35. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 907.042830] env[62066]: DEBUG nova.network.neutron [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Updating instance_info_cache with network_info: [{"id": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "address": "fa:16:3e:3c:0e:e6", "network": {"id": "49e83db7-0b32-4791-8934-d49a4300a97e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1196786879-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "33b756ddd30f4cb0b917fad171213266", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1a44472-08", "ovs_interfaceid": "d1a44472-08ab-45ac-94eb-3a8b9bb7ee35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.068910] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.102461] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.102722] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d50f0505-66f4-412f-9744-25e0ea96277a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.102878] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a8dd7483-0588-4f60-9504-20de799e69f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.103013] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.103139] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 831c0bed-8a41-4672-aa08-d60b64f365b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.103280] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a2251e56-2787-412d-89c9-eef111ee6d2b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.103413] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f24f8067-07b2-4941-8464-c30b58983ec1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.103563] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 31e50f97-f873-44ad-9923-67923cdb8d3a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.103721] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.103849] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ecf9fd16-82c7-4bea-b6a9-7262e75effef is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 907.103964] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f578eda3-3d81-418b-bcb7-0a954835ed72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.104097] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance cbf3c137-cc0f-42b7-96fb-2e1956e49b51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.105255] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance b2afea5a-2390-40b0-b860-c50f71ebac94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.105402] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 7d7743d0-acf5-426e-b409-c00dca895048 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.105662] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 00f44ecb-768a-4db3-b229-27bb6f27ede1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.105662] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d8aa9d38-0a2c-4036-9574-a85d2ed83fb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.105778] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f82c8bf2-24f3-4732-a56d-1bddaef14158 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 907.110893] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Successfully created port: 8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.325742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.326040] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.326933] env[62066]: DEBUG nova.objects.instance [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.365607] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528fd9a5-97c3-bfbd-ef07-cc41724bd4ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009175} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.366685] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06217408-071b-43d3-8224-9fe528093bd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.372332] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 907.372332] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52da903c-6d80-8dc6-ed20-1d6dfca6bf90" [ 907.372332] env[62066]: _type = "Task" [ 907.372332] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.380533] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52da903c-6d80-8dc6-ed20-1d6dfca6bf90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.545774] env[62066]: DEBUG oslo_concurrency.lockutils [req-111c29c7-4de9-4fa7-a634-8c98eafdd3bb req-2d7af857-9df3-40c6-8a03-58a111e260e7 service nova] Releasing lock "refresh_cache-d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.609310] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 4983ece5-bb4b-43cb-a928-324b2f31030e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 907.609534] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (cd4c0e36-9c88-4f73-a93c-1ff383ed97c4): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 907.804687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.805011] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.805279] env[62066]: INFO nova.compute.manager [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Attaching volume 2ac4db90-8817-40a5-a537-07020e6d7eea to /dev/sdb [ 907.835950] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59189062-8d24-4d96-80bd-8d238116f50e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.842755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d43572a-9ad1-4c77-927a-7fc974edb641 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.856175] env[62066]: DEBUG nova.virt.block_device [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating existing volume attachment record: a5fecfdd-6d30-4bfa-b271-fb9f44037f8c {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 907.882824] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52da903c-6d80-8dc6-ed20-1d6dfca6bf90, 'name': SearchDatastore_Task, 'duration_secs': 0.008901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.883078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.883361] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] d8aa9d38-0a2c-4036-9574-a85d2ed83fb2/d8aa9d38-0a2c-4036-9574-a85d2ed83fb2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.883647] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16ede7c1-b321-41d9-b2f3-ded143e2615d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.891823] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 907.891823] env[62066]: value = "task-1156407" [ 907.891823] env[62066]: _type = "Task" [ 907.891823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.904128] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156407, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.925839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.926156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.070303] env[62066]: DEBUG nova.objects.instance [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.079469] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.111149] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.111316] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.111548] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.111774] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.111941] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.112339] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.112690] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.112879] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.113042] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.113265] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.113501] env[62066]: DEBUG nova.virt.hardware [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.114309] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 908.114821] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 908.114895] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 908.118244] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eadb3ff-3e53-4453-b240-b2b1e6667e48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.129826] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68526c0f-c026-4366-8acf-d83f5ccfa57f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.404104] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156407, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458103} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.406772] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] d8aa9d38-0a2c-4036-9574-a85d2ed83fb2/d8aa9d38-0a2c-4036-9574-a85d2ed83fb2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.406997] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.408339] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73513a98-ed76-4255-813c-2daad58bdc96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.418882] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 908.418882] env[62066]: value = "task-1156411" [ 908.418882] env[62066]: _type = "Task" [ 908.418882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.429381] env[62066]: INFO nova.compute.manager [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Detaching volume c902f613-58dc-4ab2-9b93-fe59b6b4cee2 [ 908.435108] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156411, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.442947] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ace112-a489-43cc-82f3-e4d330491b71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.452471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f89151-ef53-41e4-b6b3-665315bd5579 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.487169] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33a7150-da1f-4c55-b123-9876e9b086f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.497844] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58620b84-54f2-4569-ba2f-c10d191e19da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.511700] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.573524] env[62066]: DEBUG nova.objects.base [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 908.573788] env[62066]: DEBUG nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.647766] env[62066]: DEBUG nova.policy [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.888154] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Successfully updated port: 8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.933890] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156411, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056148} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.934288] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.935097] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c77a51d-f91b-4498-a567-38d9209cdbac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.965210] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] d8aa9d38-0a2c-4036-9574-a85d2ed83fb2/d8aa9d38-0a2c-4036-9574-a85d2ed83fb2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.972179] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3acd30eb-a054-4ab0-8cfc-ae3311a2278f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.990054] env[62066]: DEBUG nova.compute.manager [req-52c7cc00-4fcc-47e4-b36c-544fa49cb9e6 req-bd2fe9c6-6c8e-417c-a846-9c3da9de5a4e service nova] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Received event network-vif-deleted-3a5c1815-3b73-48c3-af45-080c8bbb02dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.990054] env[62066]: DEBUG nova.compute.manager [req-52c7cc00-4fcc-47e4-b36c-544fa49cb9e6 req-bd2fe9c6-6c8e-417c-a846-9c3da9de5a4e service nova] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Received event network-vif-deleted-3ea61126-db59-442a-91ed-a46b37e22462 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.997408] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 908.997408] env[62066]: value = "task-1156413" [ 908.997408] env[62066]: _type = "Task" [ 908.997408] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.010844] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156413, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.014836] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.029694] env[62066]: DEBUG nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Successfully created port: 38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.392886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.393257] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquired lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.393257] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.506991] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156413, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.520291] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 909.520694] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.467s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.520826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.518s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.522377] env[62066]: INFO nova.compute.claims [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.938249] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.010526] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156413, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.119141] env[62066]: DEBUG nova.network.neutron [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updating instance_info_cache with network_info: [{"id": "8cbada80-ab13-439c-9a90-eea92446a751", "address": "fa:16:3e:91:de:39", "network": {"id": "ba0b5add-84aa-416d-ac94-4ca1b048e258", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1395858552-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcef20dbc27940be90f355cc136bd436", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5291d0-ee0f-4d70-b2ae-ab6879a67b08", "external-id": "nsx-vlan-transportzone-597", "segmentation_id": 597, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cbada80-ab", "ovs_interfaceid": "8cbada80-ab13-439c-9a90-eea92446a751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.170230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "f578eda3-3d81-418b-bcb7-0a954835ed72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.170511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.170759] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.170901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.171084] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.173753] env[62066]: INFO nova.compute.manager [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Terminating instance [ 910.175570] env[62066]: DEBUG nova.compute.manager [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.175779] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.176611] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e073a32-5da1-4ae4-a937-cbe7be47e51e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.184016] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.184337] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee279964-9c3c-4bb8-87e1-2daa859b6336 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.190675] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 910.190675] env[62066]: value = "task-1156415" [ 910.190675] env[62066]: _type = "Task" [ 910.190675] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.199450] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156415, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.328283] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.328615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.328885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.329111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.329378] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.331625] env[62066]: INFO nova.compute.manager [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Terminating instance [ 910.333837] env[62066]: DEBUG nova.compute.manager [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 910.334043] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.334862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc98b91-db40-4e3a-9009-0946f54e3042 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.342034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.342659] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f4c0055-6fa7-46c0-87e1-106480bbaa16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.349044] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 910.349044] env[62066]: value = "task-1156416" [ 910.349044] env[62066]: _type = "Task" [ 910.349044] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.356842] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.508621] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156413, 'name': ReconfigVM_Task, 'duration_secs': 1.088509} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.508945] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Reconfigured VM instance instance-00000058 to attach disk [datastore1] d8aa9d38-0a2c-4036-9574-a85d2ed83fb2/d8aa9d38-0a2c-4036-9574-a85d2ed83fb2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.509621] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b1c3dff-b901-4f8b-937f-bc6215616740 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.516561] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 910.516561] env[62066]: value = "task-1156417" [ 910.516561] env[62066]: _type = "Task" [ 910.516561] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.524476] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156417, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.525923] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 910.526125] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 910.526268] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 910.621972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Releasing lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.622361] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Instance network_info: |[{"id": "8cbada80-ab13-439c-9a90-eea92446a751", "address": "fa:16:3e:91:de:39", "network": {"id": "ba0b5add-84aa-416d-ac94-4ca1b048e258", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1395858552-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcef20dbc27940be90f355cc136bd436", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5291d0-ee0f-4d70-b2ae-ab6879a67b08", "external-id": "nsx-vlan-transportzone-597", "segmentation_id": 597, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cbada80-ab", "ovs_interfaceid": "8cbada80-ab13-439c-9a90-eea92446a751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.622780] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:de:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b5291d0-ee0f-4d70-b2ae-ab6879a67b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cbada80-ab13-439c-9a90-eea92446a751', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.630245] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Creating folder: Project (fcef20dbc27940be90f355cc136bd436). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.632655] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbcf0878-6650-4c30-8756-88659ceacdb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.642477] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Created folder: Project (fcef20dbc27940be90f355cc136bd436) in parent group-v251573. [ 910.642674] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Creating folder: Instances. Parent ref: group-v251736. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.644931] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5d103e6-dad0-4133-9e46-6a8057f64582 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.653661] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Created folder: Instances in parent group-v251736. [ 910.653896] env[62066]: DEBUG oslo.service.loopingcall [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.654104] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.654303] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84e49995-855d-4fb9-8d15-861029900030 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.674751] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.674751] env[62066]: value = "task-1156420" [ 910.674751] env[62066]: _type = "Task" [ 910.674751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.684452] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156420, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.700200] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156415, 'name': PowerOffVM_Task, 'duration_secs': 0.159478} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.700452] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.700660] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.700912] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f83e7b7-f866-4223-bc03-36abf4893f24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.768143] env[62066]: DEBUG nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Successfully updated port: 38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.773535] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.773889] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.774060] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleting the datastore file [datastore1] f578eda3-3d81-418b-bcb7-0a954835ed72 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.774281] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-077fba86-11d4-4623-bb30-45f5e2f03f01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.781183] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 910.781183] env[62066]: value = "task-1156422" [ 910.781183] env[62066]: _type = "Task" [ 910.781183] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.794326] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.827899] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56848716-e151-4194-bc3b-d11fc1596d9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.835551] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972d502b-21eb-44a7-a141-d9f2adc1ad52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.875146] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0896f0-d4d9-4e08-9f4d-e7546a31e3d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.882522] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156416, 'name': PowerOffVM_Task, 'duration_secs': 0.214248} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.884716] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.884942] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.885224] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5b929ed-7b14-420c-86f7-4c262df6fd56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.887719] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618a8c36-7798-4b0a-9856-fe19d734d67b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.901625] env[62066]: DEBUG nova.compute.provider_tree [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.956877] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.957177] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.957371] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleting the datastore file [datastore1] cbf3c137-cc0f-42b7-96fb-2e1956e49b51 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.957657] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f90fda2b-992b-4c94-91f0-49c5f3a1e84f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.964363] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for the task: (returnval){ [ 910.964363] env[62066]: value = "task-1156424" [ 910.964363] env[62066]: _type = "Task" [ 910.964363] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.972709] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.030601] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156417, 'name': Rename_Task, 'duration_secs': 0.27431} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.032590] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.035190] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Received event network-changed-f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.035758] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Refreshing instance network info cache due to event network-changed-f0fae426-1b77-435b-98a6-72d487ff589f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.035758] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquiring lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.035969] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquired lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.036207] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Refreshing network info cache for port f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.038120] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee24f3ee-5a5e-4ffa-87db-8338cef3c800 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.046092] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 911.046092] env[62066]: value = "task-1156425" [ 911.046092] env[62066]: _type = "Task" [ 911.046092] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.060476] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156425, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.144620] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 911.144911] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251733', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'name': 'volume-cd885df2-f1e2-4754-bb18-08a36f7f908d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d7743d0-acf5-426e-b409-c00dca895048', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'serial': 'cd885df2-f1e2-4754-bb18-08a36f7f908d'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 911.145839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e988e2ba-55ee-4e61-bce2-fcd39be9a193 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.164151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1521a9c-0133-4cbc-a33f-e33b988b6557 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.191516] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] volume-cd885df2-f1e2-4754-bb18-08a36f7f908d/volume-cd885df2-f1e2-4754-bb18-08a36f7f908d.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.198418] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc2c2523-c2b7-41d6-a55d-8bf873e6fcaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.220023] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156420, 'name': CreateVM_Task, 'duration_secs': 0.384107} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.221442] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.221941] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 911.221941] env[62066]: value = "task-1156426" [ 911.221941] env[62066]: _type = "Task" [ 911.221941] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.222747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.222964] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.223376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.223742] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb0679e6-b0a6-43a6-b360-04c41ffad93b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.231770] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 911.231770] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52496f2a-a4c8-f12c-f958-7264470c3c07" [ 911.231770] env[62066]: _type = "Task" [ 911.231770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.235127] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156426, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.239514] env[62066]: INFO nova.virt.block_device [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Attempting to driver detach volume c902f613-58dc-4ab2-9b93-fe59b6b4cee2 from mountpoint /dev/sdb [ 911.239785] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 911.239983] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251705', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'name': 'volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '831c0bed-8a41-4672-aa08-d60b64f365b8', 'attached_at': '', 'detached_at': '', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'serial': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 911.240789] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8910bf-e9e1-46a5-a6b5-1b130d6455ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.246930] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52496f2a-a4c8-f12c-f958-7264470c3c07, 'name': SearchDatastore_Task, 'duration_secs': 0.009325} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.247615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.247876] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.248145] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.248297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.248480] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.248743] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03a0cc85-2ce3-4974-b11a-d230ed75b3d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.267146] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec27dbd3-5b63-46e2-a2da-96bdb5f7295b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.273961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.274171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.274400] env[62066]: DEBUG nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.277038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1110da30-1954-4992-9911-bdd006669e90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.280017] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.280222] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.281664] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111908de-95ba-4097-8af7-d84a698a473b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.308677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812e6bbd-1bc3-4b9e-a4e6-bbf2ab7d94a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.311552] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 911.311552] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52973519-2bf6-1f67-a415-9eddfe3a0830" [ 911.311552] env[62066]: _type = "Task" [ 911.311552] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.315356] env[62066]: DEBUG oslo_vmware.api [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126148} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.318512] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.318710] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.318893] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.319082] env[62066]: INFO nova.compute.manager [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Took 1.14 seconds to destroy the instance on the hypervisor. [ 911.319319] env[62066]: DEBUG oslo.service.loopingcall [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.330058] env[62066]: DEBUG nova.compute.manager [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.330236] env[62066]: DEBUG nova.network.neutron [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.331826] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] The volume has not been displaced from its original location: [datastore2] volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2/volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 911.337047] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 911.337851] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48920dfd-0b41-43ac-b7b8-9930398a2e31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.354691] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52973519-2bf6-1f67-a415-9eddfe3a0830, 'name': SearchDatastore_Task, 'duration_secs': 0.008726} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.355778] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58abfc6f-a784-4b81-8678-036c05fe0a9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.359744] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 911.359744] env[62066]: value = "task-1156427" [ 911.359744] env[62066]: _type = "Task" [ 911.359744] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.360976] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 911.360976] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f24d31-b50b-8909-fc72-734acb49a3fd" [ 911.360976] env[62066]: _type = "Task" [ 911.360976] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.371692] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156427, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.374649] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f24d31-b50b-8909-fc72-734acb49a3fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.405079] env[62066]: DEBUG nova.scheduler.client.report [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.475832] env[62066]: DEBUG oslo_vmware.api [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Task: {'id': task-1156424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121195} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.476112] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.476299] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.476553] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.476662] env[62066]: INFO nova.compute.manager [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Took 1.14 seconds to destroy the instance on the hypervisor. [ 911.476904] env[62066]: DEBUG oslo.service.loopingcall [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.477110] env[62066]: DEBUG nova.compute.manager [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.477212] env[62066]: DEBUG nova.network.neutron [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 911.533571] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 911.533988] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.534072] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.534247] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.534420] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.534580] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.534746] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 911.535100] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 911.557507] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156425, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.734016] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156426, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.790553] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updated VIF entry in instance network info cache for port f0fae426-1b77-435b-98a6-72d487ff589f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.790942] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating instance_info_cache with network_info: [{"id": "f0fae426-1b77-435b-98a6-72d487ff589f", "address": "fa:16:3e:8c:82:6e", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fae426-1b", "ovs_interfaceid": "f0fae426-1b77-435b-98a6-72d487ff589f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.841737] env[62066]: WARNING nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 911.877010] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f24d31-b50b-8909-fc72-734acb49a3fd, 'name': SearchDatastore_Task, 'duration_secs': 0.015314} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.881368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.881654] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f82c8bf2-24f3-4732-a56d-1bddaef14158/f82c8bf2-24f3-4732-a56d-1bddaef14158.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.882332] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156427, 'name': ReconfigVM_Task, 'duration_secs': 0.292639} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.882538] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21dff64c-fc32-4a2a-9fc7-fc916c7b80bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.884549] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 911.889110] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-657d6a7c-046a-43a8-b267-7943bb843025 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.907330] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 911.907330] env[62066]: value = "task-1156428" [ 911.907330] env[62066]: _type = "Task" [ 911.907330] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.908825] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 911.908825] env[62066]: value = "task-1156429" [ 911.908825] env[62066]: _type = "Task" [ 911.908825] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.912242] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.912810] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.918296] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.530s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.918495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.921944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.051s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.922062] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.923986] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.450s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.924194] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.924372] env[62066]: INFO nova.compute.manager [None req-2df5607a-4be6-4881-a99c-819bb69adea8 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Successfully reverted task state from rebuilding on failure for instance. [ 911.930277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.137s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.930437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.932422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.854s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.934155] env[62066]: INFO nova.compute.claims [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.945116] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.948589] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156429, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.964029] env[62066]: INFO nova.scheduler.client.report [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Deleted allocations for instance a2251e56-2787-412d-89c9-eef111ee6d2b [ 911.972057] env[62066]: INFO nova.scheduler.client.report [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted allocations for instance ccb9f50f-dcc3-4d81-944e-d70803185ae1 [ 912.059997] env[62066]: DEBUG oslo_vmware.api [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156425, 'name': PowerOnVM_Task, 'duration_secs': 0.839751} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.060559] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.060701] env[62066]: INFO nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Took 10.26 seconds to spawn the instance on the hypervisor. [ 912.060885] env[62066]: DEBUG nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.061667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255b550b-3b77-4375-b23b-90eb6db5cc9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.206246] env[62066]: DEBUG nova.network.neutron [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.236326] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156426, 'name': ReconfigVM_Task, 'duration_secs': 0.5556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.239317] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfigured VM instance instance-00000056 to attach disk [datastore1] volume-cd885df2-f1e2-4754-bb18-08a36f7f908d/volume-cd885df2-f1e2-4754-bb18-08a36f7f908d.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.246972] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39bde534-e00a-4f09-a3bd-3da5a8b93476 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.265388] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 912.265388] env[62066]: value = "task-1156430" [ 912.265388] env[62066]: _type = "Task" [ 912.265388] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.275951] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.296947] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Releasing lock "refresh_cache-00f44ecb-768a-4db3-b229-27bb6f27ede1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.297372] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Received event network-vif-plugged-8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.297701] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquiring lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.298016] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.298301] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.298521] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] No waiting events found dispatching network-vif-plugged-8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.298747] env[62066]: WARNING nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Received unexpected event network-vif-plugged-8cbada80-ab13-439c-9a90-eea92446a751 for instance with vm_state building and task_state spawning. [ 912.299110] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Received event network-changed-8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.299110] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Refreshing instance network info cache due to event network-changed-8cbada80-ab13-439c-9a90-eea92446a751. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.299307] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquiring lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.299452] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquired lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.299611] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Refreshing network info cache for port 8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.310011] env[62066]: DEBUG nova.network.neutron [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38b4765c-0887-499c-b63b-895e1168aa5d", "address": "fa:16:3e:a7:45:38", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38b4765c-08", "ovs_interfaceid": "38b4765c-0887-499c-b63b-895e1168aa5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.336893] env[62066]: DEBUG nova.network.neutron [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.409465] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 912.409465] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251735', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'name': 'volume-2ac4db90-8817-40a5-a537-07020e6d7eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31e50f97-f873-44ad-9923-67923cdb8d3a', 'attached_at': '', 'detached_at': '', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'serial': '2ac4db90-8817-40a5-a537-07020e6d7eea'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 912.409777] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98b0a1b-7347-47a5-93ec-b5eb8c4231e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.426000] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156428, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.441328] env[62066]: DEBUG oslo_vmware.api [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156429, 'name': ReconfigVM_Task, 'duration_secs': 0.137301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.443167] env[62066]: DEBUG nova.compute.utils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 912.446731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7442a9d-60bf-41c4-ad38-d6c1d0185ea0 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.587s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.448058] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251705', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'name': 'volume-c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '831c0bed-8a41-4672-aa08-d60b64f365b8', 'attached_at': '', 'detached_at': '', 'volume_id': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2', 'serial': 'c902f613-58dc-4ab2-9b93-fe59b6b4cee2'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 912.450542] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.450884] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.453889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c84f6e-ef14-4682-9402-e7c33e8eabf1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.455957] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 10.186s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.456150] env[62066]: INFO nova.compute.manager [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Unshelving [ 913.242026] env[62066]: INFO nova.compute.manager [-] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Took 1.91 seconds to deallocate network for instance. [ 913.242026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.242026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.242026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.242026] env[62066]: INFO nova.compute.manager [-] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Took 1.76 seconds to deallocate network for instance. [ 913.243236] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.254414] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] volume-2ac4db90-8817-40a5-a537-07020e6d7eea/volume-2ac4db90-8817-40a5-a537-07020e6d7eea.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.255157] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feba5df5-1e22-4398-b694-46734b960830 tempest-AttachVolumeShelveTestJSON-1752912476 tempest-AttachVolumeShelveTestJSON-1752912476-project-member] Lock "a2251e56-2787-412d-89c9-eef111ee6d2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.893s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.259260] env[62066]: DEBUG nova.policy [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.261032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0ba10cbc-fae6-4037-93dd-84c3ac01ba77 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ccb9f50f-dcc3-4d81-944e-d70803185ae1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.036s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.262852] env[62066]: DEBUG nova.objects.instance [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'flavor' on Instance uuid 831c0bed-8a41-4672-aa08-d60b64f365b8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.271606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6326582d-7223-42c6-942d-474975bef86a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.278235] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-200f59f3-d3f4-4861-8e15-4c0f385eb449 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.293258] env[62066]: INFO nova.compute.manager [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Took 22.50 seconds to build instance. [ 913.319449] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156428, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518508} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.320030] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.320267] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.320423] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.320615] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.320831] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.320910] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.321135] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.321297] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.321465] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.321663] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.321808] env[62066]: DEBUG nova.virt.hardware [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.328870] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 913.336617] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] f82c8bf2-24f3-4732-a56d-1bddaef14158/f82c8bf2-24f3-4732-a56d-1bddaef14158.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.336870] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.337479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3696c3b-28db-471d-a6cb-f133bcde0851 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.350121] env[62066]: DEBUG oslo_vmware.api [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156430, 'name': ReconfigVM_Task, 'duration_secs': 0.182456} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.350245] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 913.350245] env[62066]: value = "task-1156431" [ 913.350245] env[62066]: _type = "Task" [ 913.350245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.353247] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f512c87-5414-415a-8cdc-38fc1c89f4ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.356785] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251733', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'name': 'volume-cd885df2-f1e2-4754-bb18-08a36f7f908d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d7743d0-acf5-426e-b409-c00dca895048', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'serial': 'cd885df2-f1e2-4754-bb18-08a36f7f908d'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 913.366329] env[62066]: DEBUG oslo_vmware.api [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 913.366329] env[62066]: value = "task-1156432" [ 913.366329] env[62066]: _type = "Task" [ 913.366329] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.368333] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 913.368333] env[62066]: value = "task-1156433" [ 913.368333] env[62066]: _type = "Task" [ 913.368333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.375993] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.396029] env[62066]: DEBUG oslo_vmware.api [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156432, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.396219] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156433, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.491633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.491941] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.630781] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ac427f-ab82-41a6-aae3-bc6a6d3062fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.638402] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66b03b1-681f-4801-aa66-8e54821eedbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.675450] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updated VIF entry in instance network info cache for port 8cbada80-ab13-439c-9a90-eea92446a751. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.675814] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updating instance_info_cache with network_info: [{"id": "8cbada80-ab13-439c-9a90-eea92446a751", "address": "fa:16:3e:91:de:39", "network": {"id": "ba0b5add-84aa-416d-ac94-4ca1b048e258", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1395858552-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcef20dbc27940be90f355cc136bd436", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5291d0-ee0f-4d70-b2ae-ab6879a67b08", "external-id": "nsx-vlan-transportzone-597", "segmentation_id": 597, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cbada80-ab", "ovs_interfaceid": "8cbada80-ab13-439c-9a90-eea92446a751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.678878] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4ba68e-2082-40d7-9ca3-b7c2781db44b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.688448] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4afb2e6-41e9-415f-8db5-3cbd9029b4a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.707311] env[62066]: DEBUG nova.compute.provider_tree [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.777724] env[62066]: DEBUG nova.compute.manager [req-cdbe912e-1980-4b2e-bf21-587edf595127 req-c7315694-c34e-4738-87f9-de3609a45662 service nova] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Received event network-vif-deleted-55ac0005-052f-46d1-9bf5-8807457b4fa8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.777963] env[62066]: DEBUG nova.compute.manager [req-cdbe912e-1980-4b2e-bf21-587edf595127 req-c7315694-c34e-4738-87f9-de3609a45662 service nova] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Received event network-vif-deleted-d9083d7c-317d-46e6-bf0f-eb924b695754 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.796552] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Successfully created port: 30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.798975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bf2d176-47b5-4794-84d4-74a4d07fed48 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.015s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.800606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.871057] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.879298] env[62066]: DEBUG oslo_vmware.api [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.888970] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156433, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.889191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.889930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7256f4a4-28bd-49ac-985d-b9f68a0ed216 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.893437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.914157] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] f82c8bf2-24f3-4732-a56d-1bddaef14158/f82c8bf2-24f3-4732-a56d-1bddaef14158.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.915488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.917580] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ad4725e-974f-4a9b-bcdf-ba043a0a0e98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.932290] env[62066]: DEBUG nova.objects.instance [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'flavor' on Instance uuid 7d7743d0-acf5-426e-b409-c00dca895048 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.939561] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 913.939561] env[62066]: value = "task-1156434" [ 913.939561] env[62066]: _type = "Task" [ 913.939561] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.947667] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156434, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.994447] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.184592] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Releasing lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.184949] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-plugged-38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.185251] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.185600] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.185821] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.186111] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] No waiting events found dispatching network-vif-plugged-38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.186335] env[62066]: WARNING nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received unexpected event network-vif-plugged-38b4765c-0887-499c-b63b-895e1168aa5d for instance with vm_state active and task_state None. [ 914.186566] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-changed-38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.186782] env[62066]: DEBUG nova.compute.manager [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing instance network info cache due to event network-changed-38b4765c-0887-499c-b63b-895e1168aa5d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.187056] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.187574] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.187574] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing network info cache for port 38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.210640] env[62066]: DEBUG nova.scheduler.client.report [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.267717] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.293933] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.294734] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.294734] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.294870] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.294987] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.295163] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.295402] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.296723] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.296723] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.296723] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.296723] env[62066]: DEBUG nova.virt.hardware [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.297380] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f9738d-0596-42ab-8b3f-7a42d73346ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.300959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0368a460-6b47-4b80-b05d-49ff7461479c tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 6.374s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.308536] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22d3be5-ddc4-4df3-8acc-a845c1c04bf2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.373486] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156431, 'name': ReconfigVM_Task, 'duration_secs': 0.826303} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.376930] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfigured VM instance instance-0000004e to attach disk [datastore1] volume-2ac4db90-8817-40a5-a537-07020e6d7eea/volume-2ac4db90-8817-40a5-a537-07020e6d7eea.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.381760] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63cd14d1-d1a3-4c0d-9971-7380c617d458 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.398508] env[62066]: DEBUG oslo_vmware.api [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156432, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.400080] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 914.400080] env[62066]: value = "task-1156435" [ 914.400080] env[62066]: _type = "Task" [ 914.400080] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.410091] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156435, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.437579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77d9ab6e-5d45-4d1e-b41d-7a3bb1362691 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.908s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.449598] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156434, 'name': ReconfigVM_Task, 'duration_secs': 0.419481} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.449598] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Reconfigured VM instance instance-00000059 to attach disk [datastore1] f82c8bf2-24f3-4732-a56d-1bddaef14158/f82c8bf2-24f3-4732-a56d-1bddaef14158.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.450328] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64eaffe9-9022-4c68-a45b-74f5408311dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.457910] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 914.457910] env[62066]: value = "task-1156436" [ 914.457910] env[62066]: _type = "Task" [ 914.457910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.465916] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156436, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.517843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.719867] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.783s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.719867] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.719867] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.468s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.720094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.722475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.212s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.722945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.727645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.925s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.727645] env[62066]: DEBUG nova.objects.instance [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lazy-loading 'resources' on Instance uuid f578eda3-3d81-418b-bcb7-0a954835ed72 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.754201] env[62066]: DEBUG nova.compute.manager [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.755524] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f96fa72-047f-4979-90fc-ea1934b36991 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.759324] env[62066]: INFO nova.scheduler.client.report [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocations for instance ecf9fd16-82c7-4bea-b6a9-7262e75effef [ 914.885017] env[62066]: DEBUG oslo_vmware.api [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156432, 'name': ReconfigVM_Task, 'duration_secs': 1.141818} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.885743] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.886056] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 914.912063] env[62066]: DEBUG oslo_vmware.api [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156435, 'name': ReconfigVM_Task, 'duration_secs': 0.289676} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.912370] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251735', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'name': 'volume-2ac4db90-8817-40a5-a537-07020e6d7eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31e50f97-f873-44ad-9923-67923cdb8d3a', 'attached_at': '', 'detached_at': '', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'serial': '2ac4db90-8817-40a5-a537-07020e6d7eea'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 914.963476] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updated VIF entry in instance network info cache for port 38b4765c-0887-499c-b63b-895e1168aa5d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 914.963954] env[62066]: DEBUG nova.network.neutron [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38b4765c-0887-499c-b63b-895e1168aa5d", "address": "fa:16:3e:a7:45:38", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38b4765c-08", "ovs_interfaceid": "38b4765c-0887-499c-b63b-895e1168aa5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.970564] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156436, 'name': Rename_Task, 'duration_secs': 0.199361} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.970848] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.971114] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-671ca8ce-ca43-4aa6-ac0b-a92be28c49b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.977195] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 914.977195] env[62066]: value = "task-1156437" [ 914.977195] env[62066]: _type = "Task" [ 914.977195] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.986396] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.214522] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.214826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.215060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7d7743d0-acf5-426e-b409-c00dca895048-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.215256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.215451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.218827] env[62066]: INFO nova.compute.manager [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Terminating instance [ 915.221053] env[62066]: DEBUG nova.compute.manager [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.221619] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.221619] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8a7c476-c78c-4461-a5c6-c6b81250a128 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.229201] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 915.229201] env[62066]: value = "task-1156438" [ 915.229201] env[62066]: _type = "Task" [ 915.229201] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.231120] env[62066]: DEBUG nova.compute.utils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 915.242033] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.242033] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 915.250517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3144e2f3-8e0c-4768-8f66-f20de66220f4 tempest-ServerActionsV293TestJSON-1220992018 tempest-ServerActionsV293TestJSON-1220992018-project-member] Lock "753f585a-22d7-4eeb-8580-4e3a68b5fd72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.786s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.258936] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.271342] env[62066]: INFO nova.compute.manager [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] instance snapshotting [ 915.278865] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4f98992-52f2-450f-a458-8fe34f2d4a53 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "ecf9fd16-82c7-4bea-b6a9-7262e75effef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.900s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.283722] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60aa729d-e06e-4c9c-a36e-8fda9b5a8c1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.324123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e319820a-7986-4bdd-961e-53bcfbed8e92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.391067] env[62066]: DEBUG oslo_concurrency.lockutils [None req-db1d34c3-206e-4cea-bf15-bd8de420f3ae tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.065s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.468207] env[62066]: DEBUG oslo_concurrency.lockutils [req-cb4a2bf4-352d-4de7-bf11-70125dde4f4f req-1c67b8cd-f1ac-4937-ace0-02d0991cb583 service nova] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.491490] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156437, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.534792] env[62066]: DEBUG nova.policy [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85a0637645c443a9d9e6028053c0450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '350ba3c5676a4dd0a018900e7237a5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.610516] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4241fbb8-3aeb-41a6-8575-0078299e463b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.620229] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35c4eb5-3ec2-4c19-b650-ca8c602cc58f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.654373] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a6d83d-48c7-42b2-afb6-bdd5274f917c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.662440] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04308727-4ea6-4d05-b70b-b542a29d30dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.676507] env[62066]: DEBUG nova.compute.provider_tree [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.734628] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.740470] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156438, 'name': PowerOffVM_Task, 'duration_secs': 0.203982} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.741053] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.741277] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 915.741481] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251733', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'name': 'volume-cd885df2-f1e2-4754-bb18-08a36f7f908d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d7743d0-acf5-426e-b409-c00dca895048', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'serial': 'cd885df2-f1e2-4754-bb18-08a36f7f908d'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 915.742336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfaaead-b4d1-4826-846f-c65fac77c2ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.765982] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Successfully updated port: 30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.765982] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4daaaac-4182-4059-b4d1-3ea8b40b9131 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.773330] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd46a485-4b09-4e3a-bd90-acc7ee33aa58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.795411] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a59791-32df-412b-9247-6f6a57be7ae0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.811497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.811768] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.811967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.812233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.812333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.814115] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] The volume has not been displaced from its original location: [datastore1] volume-cd885df2-f1e2-4754-bb18-08a36f7f908d/volume-cd885df2-f1e2-4754-bb18-08a36f7f908d.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 915.819409] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 915.819991] env[62066]: INFO nova.compute.manager [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Terminating instance [ 915.821816] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53f2293d-2b69-4d9e-b1dc-71808911c42e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.835325] env[62066]: DEBUG nova.compute.manager [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.835493] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.836593] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60755b07-afc1-43ca-bc48-ade137c772b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.840372] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 915.840898] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af5f6e9d-d8ab-41d0-86bf-b817cdc1ea1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.847121] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 915.847121] env[62066]: value = "task-1156439" [ 915.847121] env[62066]: _type = "Task" [ 915.847121] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.849206] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.853037] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b25c624-5879-4dad-aea2-5e1b1216fe37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.854451] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 915.854451] env[62066]: value = "task-1156440" [ 915.854451] env[62066]: _type = "Task" [ 915.854451] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.861375] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156439, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.863081] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 915.863081] env[62066]: value = "task-1156441" [ 915.863081] env[62066]: _type = "Task" [ 915.863081] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.870119] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156440, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.875596] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.962717] env[62066]: DEBUG nova.objects.instance [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.992651] env[62066]: DEBUG oslo_vmware.api [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156437, 'name': PowerOnVM_Task, 'duration_secs': 0.665434} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.992961] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.993856] env[62066]: INFO nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Took 7.91 seconds to spawn the instance on the hypervisor. [ 915.993856] env[62066]: DEBUG nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.994222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0881bc-c5c2-46f9-8d23-ecffb77ac688 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.122083] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Successfully created port: aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.179285] env[62066]: DEBUG nova.scheduler.client.report [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.269898] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.270187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.270261] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.361593] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156439, 'name': ReconfigVM_Task, 'duration_secs': 0.224355} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.362348] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 916.369914] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d46d8cf0-213d-4871-b161-62cf90579c17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.380177] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156440, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.389474] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156441, 'name': PowerOffVM_Task, 'duration_secs': 0.157536} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.391370] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.391554] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.391902] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 916.391902] env[62066]: value = "task-1156442" [ 916.391902] env[62066]: _type = "Task" [ 916.391902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.392120] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbf0fbe3-d887-4146-9a05-c895219dff5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.402266] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156442, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.468656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc48dc9f-62f7-43d5-af31-d3449ded266a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.663s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.476353] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.476672] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.477137] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleting the datastore file [datastore2] 831c0bed-8a41-4672-aa08-d60b64f365b8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.481312] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aeff6958-b148-429f-b551-0c30fa639b71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.484035] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 916.484035] env[62066]: value = "task-1156444" [ 916.484035] env[62066]: _type = "Task" [ 916.484035] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.492726] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.503422] env[62066]: DEBUG nova.compute.manager [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Received event network-vif-plugged-30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.503635] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] Acquiring lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.503857] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.504032] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.504206] env[62066]: DEBUG nova.compute.manager [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] No waiting events found dispatching network-vif-plugged-30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.504374] env[62066]: WARNING nova.compute.manager [req-1c5bd314-21b8-464b-897d-08ab963f404a req-bb418c4c-20dc-4cdc-a909-28db5ae8f87f service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Received unexpected event network-vif-plugged-30527f31-db37-41fe-931c-15eb75cc3794 for instance with vm_state building and task_state spawning. [ 916.511236] env[62066]: DEBUG nova.compute.manager [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Received event network-changed-30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.511395] env[62066]: DEBUG nova.compute.manager [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Refreshing instance network info cache due to event network-changed-30527f31-db37-41fe-931c-15eb75cc3794. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.511851] env[62066]: DEBUG oslo_concurrency.lockutils [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] Acquiring lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.514556] env[62066]: INFO nova.compute.manager [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Took 23.95 seconds to build instance. [ 916.683885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.686603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.793s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.687063] env[62066]: DEBUG nova.objects.instance [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lazy-loading 'resources' on Instance uuid cbf3c137-cc0f-42b7-96fb-2e1956e49b51 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.711275] env[62066]: INFO nova.scheduler.client.report [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted allocations for instance f578eda3-3d81-418b-bcb7-0a954835ed72 [ 916.749017] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.789343] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.789697] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.790418] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.791100] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.791290] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.791452] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.791712] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.791868] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.792076] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.792984] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.792984] env[62066]: DEBUG nova.virt.hardware [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.793994] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c78ef3f-5ccf-4859-8141-5efc954854a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.807470] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1937b9-f44c-4194-be65-fa2d67bb2b16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.829495] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.867110] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156440, 'name': CreateSnapshot_Task, 'duration_secs': 0.783257} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.867416] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 916.868178] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d3a433-2d84-4cb5-b8f7-1b54895928a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.907548] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156442, 'name': ReconfigVM_Task, 'duration_secs': 0.178282} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.907838] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251733', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'name': 'volume-cd885df2-f1e2-4754-bb18-08a36f7f908d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7d7743d0-acf5-426e-b409-c00dca895048', 'attached_at': '', 'detached_at': '', 'volume_id': 'cd885df2-f1e2-4754-bb18-08a36f7f908d', 'serial': 'cd885df2-f1e2-4754-bb18-08a36f7f908d'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 916.908136] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.910194] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2663f8-011a-4106-9e22-c078d656f62b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.917581] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.918525] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3310aed7-789b-472f-b4cf-e9b1f5f692f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.977826] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.978120] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.978322] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore1] 7d7743d0-acf5-426e-b409-c00dca895048 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.978615] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7451e10a-42ea-40da-81dc-0c61b59efbc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.985371] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 916.985371] env[62066]: value = "task-1156446" [ 916.985371] env[62066]: _type = "Task" [ 916.985371] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.998365] env[62066]: DEBUG oslo_vmware.api [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145956} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.001546] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.001767] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.001950] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.002147] env[62066]: INFO nova.compute.manager [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Took 1.17 seconds to destroy the instance on the hypervisor. [ 917.002517] env[62066]: DEBUG oslo.service.loopingcall [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.002750] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.002983] env[62066]: DEBUG nova.compute.manager [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.003103] env[62066]: DEBUG nova.network.neutron [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.018615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e90754d3-c921-469a-b569-8f6d33e60986 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.465s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.110335] env[62066]: DEBUG nova.network.neutron [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Updating instance_info_cache with network_info: [{"id": "30527f31-db37-41fe-931c-15eb75cc3794", "address": "fa:16:3e:35:00:f6", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30527f31-db", "ovs_interfaceid": "30527f31-db37-41fe-931c-15eb75cc3794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.227111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cc3d054d-27e8-45bb-b64f-93aa81524e95 tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "f578eda3-3d81-418b-bcb7-0a954835ed72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.055s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.303600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.304239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.304800] env[62066]: DEBUG nova.compute.manager [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.305873] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb541c09-cf47-4474-9002-d57c84382e24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.316895] env[62066]: DEBUG nova.compute.manager [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 917.316895] env[62066]: DEBUG nova.objects.instance [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.386500] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 917.389180] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-557638e3-65b2-47a3-a093-ac2311ddec5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.400858] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 917.400858] env[62066]: value = "task-1156447" [ 917.400858] env[62066]: _type = "Task" [ 917.400858] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.413261] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156447, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.506366] env[62066]: DEBUG oslo_vmware.api [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142841} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.507894] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.508183] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.508427] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.508669] env[62066]: INFO nova.compute.manager [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Took 2.29 seconds to destroy the instance on the hypervisor. [ 917.508977] env[62066]: DEBUG oslo.service.loopingcall [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.509982] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352d6b12-6299-4a2f-88fb-e704afad6ae2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.513789] env[62066]: DEBUG nova.compute.manager [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.513789] env[62066]: DEBUG nova.network.neutron [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.521723] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d51696e-89b7-49a8-80b1-740920ac26c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.580663] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37e980b-9b4e-40de-8aea-53516f3f01cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.597739] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3024d021-2fe7-415e-88d4-7d6f9cadd04b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.618669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.619107] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Instance network_info: |[{"id": "30527f31-db37-41fe-931c-15eb75cc3794", "address": "fa:16:3e:35:00:f6", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30527f31-db", "ovs_interfaceid": "30527f31-db37-41fe-931c-15eb75cc3794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.621810] env[62066]: DEBUG nova.compute.provider_tree [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.622406] env[62066]: DEBUG oslo_concurrency.lockutils [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] Acquired lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.622754] env[62066]: DEBUG nova.network.neutron [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Refreshing network info cache for port 30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.624331] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:00:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30527f31-db37-41fe-931c-15eb75cc3794', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.634970] env[62066]: DEBUG oslo.service.loopingcall [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.636505] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.636748] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-797ba2d0-ff3d-40bc-8c7f-a17a348ee6fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.659963] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.659963] env[62066]: value = "task-1156448" [ 917.659963] env[62066]: _type = "Task" [ 917.659963] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.672025] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156448, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.822684] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.823160] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a900e926-47c4-4909-aa85-caa8b61cc468 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.830866] env[62066]: DEBUG oslo_vmware.api [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 917.830866] env[62066]: value = "task-1156449" [ 917.830866] env[62066]: _type = "Task" [ 917.830866] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.854618] env[62066]: DEBUG oslo_vmware.api [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.917741] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156447, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.140700] env[62066]: DEBUG nova.scheduler.client.report [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.175729] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156448, 'name': CreateVM_Task, 'duration_secs': 0.35549} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.175998] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.177068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.177408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.177975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.179166] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93ed9f1f-3648-4272-a395-85e25e6df631 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.188918] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 918.188918] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52910d3f-bcb1-04c9-b377-94c4d34bb8fa" [ 918.188918] env[62066]: _type = "Task" [ 918.188918] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.210628] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52910d3f-bcb1-04c9-b377-94c4d34bb8fa, 'name': SearchDatastore_Task, 'duration_secs': 0.011924} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.210628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.210628] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.210945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.211333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.211827] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.212328] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14a62462-539c-4dbb-aa69-dbc16261cda5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.225542] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.226054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.227272] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d11bb75c-4c96-4b46-ad4b-dfccae1311d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.234422] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 918.234422] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52328c88-c159-a7af-07a6-b1b1bed613c7" [ 918.234422] env[62066]: _type = "Task" [ 918.234422] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.242969] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52328c88-c159-a7af-07a6-b1b1bed613c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.341450] env[62066]: DEBUG oslo_vmware.api [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156449, 'name': PowerOffVM_Task, 'duration_secs': 0.465901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.344271] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.344547] env[62066]: DEBUG nova.compute.manager [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.345856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4d25b4-1a58-4ae9-93e5-c679955ba5a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.360530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-5fb737fe-ca04-4ac5-851d-d2d9397ae02d" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.361016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-5fb737fe-ca04-4ac5-851d-d2d9397ae02d" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.361426] env[62066]: DEBUG nova.objects.instance [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.413654] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156447, 'name': CloneVM_Task} progress is 95%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.632308] env[62066]: DEBUG nova.network.neutron [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.647356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.961s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.650125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.735s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.650455] env[62066]: DEBUG nova.objects.instance [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lazy-loading 'pci_requests' on Instance uuid 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.652782] env[62066]: DEBUG nova.network.neutron [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.677900] env[62066]: DEBUG nova.network.neutron [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Updated VIF entry in instance network info cache for port 30527f31-db37-41fe-931c-15eb75cc3794. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.678309] env[62066]: DEBUG nova.network.neutron [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Updating instance_info_cache with network_info: [{"id": "30527f31-db37-41fe-931c-15eb75cc3794", "address": "fa:16:3e:35:00:f6", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30527f31-db", "ovs_interfaceid": "30527f31-db37-41fe-931c-15eb75cc3794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.685755] env[62066]: INFO nova.scheduler.client.report [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Deleted allocations for instance cbf3c137-cc0f-42b7-96fb-2e1956e49b51 [ 918.706411] env[62066]: DEBUG nova.compute.manager [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Received event network-changed-8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.706411] env[62066]: DEBUG nova.compute.manager [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Refreshing instance network info cache due to event network-changed-8cbada80-ab13-439c-9a90-eea92446a751. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.706411] env[62066]: DEBUG oslo_concurrency.lockutils [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] Acquiring lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.706411] env[62066]: DEBUG oslo_concurrency.lockutils [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] Acquired lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.706411] env[62066]: DEBUG nova.network.neutron [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Refreshing network info cache for port 8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.750341] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52328c88-c159-a7af-07a6-b1b1bed613c7, 'name': SearchDatastore_Task, 'duration_secs': 0.018736} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.750341] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a62633a-7fdf-4e2b-80c5-a645138d6ace {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.755144] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 918.755144] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529bbb33-857e-a128-958b-0cd7a1b9d170" [ 918.755144] env[62066]: _type = "Task" [ 918.755144] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.765516] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529bbb33-857e-a128-958b-0cd7a1b9d170, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.868379] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e5cebf0-2da1-4dae-8e1f-10e7c40e7a1c tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.564s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.888806] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Successfully updated port: aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.917686] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156447, 'name': CloneVM_Task, 'duration_secs': 1.222552} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.918077] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Created linked-clone VM from snapshot [ 918.918903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaafda30-9085-4845-b538-1c2b66856865 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.931412] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Uploading image 84851c67-5ec7-44f1-906b-ef3db11ba04a {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 918.957041] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 918.957041] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f35e0c1c-a283-461c-8c06-34e82e1cd4b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.963537] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 918.963537] env[62066]: value = "task-1156450" [ 918.963537] env[62066]: _type = "Task" [ 918.963537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.972450] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156450, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.053304] env[62066]: DEBUG nova.objects.instance [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.134942] env[62066]: INFO nova.compute.manager [-] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Took 1.62 seconds to deallocate network for instance. [ 919.154429] env[62066]: INFO nova.compute.manager [-] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Took 2.15 seconds to deallocate network for instance. [ 919.156298] env[62066]: DEBUG nova.objects.instance [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lazy-loading 'numa_topology' on Instance uuid 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.180501] env[62066]: DEBUG oslo_concurrency.lockutils [req-447261ab-2624-4965-9d31-f2b4bf340779 req-f1229925-bb1d-4300-9bbd-917910f28d19 service nova] Releasing lock "refresh_cache-4983ece5-bb4b-43cb-a928-324b2f31030e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.196265] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fa4b53e6-2cad-42be-adce-010a0cd80eae tempest-ListServersNegativeTestJSON-1436849271 tempest-ListServersNegativeTestJSON-1436849271-project-member] Lock "cbf3c137-cc0f-42b7-96fb-2e1956e49b51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.867s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.267355] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529bbb33-857e-a128-958b-0cd7a1b9d170, 'name': SearchDatastore_Task, 'duration_secs': 0.021881} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.267652] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.267924] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 4983ece5-bb4b-43cb-a928-324b2f31030e/4983ece5-bb4b-43cb-a928-324b2f31030e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.268216] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea01138e-c6a3-416c-bec1-559b8ad8766f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.275901] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 919.275901] env[62066]: value = "task-1156451" [ 919.275901] env[62066]: _type = "Task" [ 919.275901] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.285340] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.392608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.392817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.393015] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.455019] env[62066]: DEBUG nova.compute.manager [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Received event network-vif-plugged-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.455296] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.455500] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.455772] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.456020] env[62066]: DEBUG nova.compute.manager [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] No waiting events found dispatching network-vif-plugged-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.456227] env[62066]: WARNING nova.compute.manager [req-1c268d75-ac72-4f76-aa5a-484e1f960212 req-87c91b87-9ffe-4f61-be90-bc2b400ac824 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Received unexpected event network-vif-plugged-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f for instance with vm_state building and task_state spawning. [ 919.482788] env[62066]: DEBUG nova.compute.manager [req-900c0803-a281-4a39-a289-d95fbdb2834f req-894acab3-09b6-466b-a847-7c732c267943 service nova] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Received event network-vif-deleted-b010e41f-8864-4417-b170-3b94f1b1b74d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.483516] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156450, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.557453] env[62066]: DEBUG nova.objects.base [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 919.557824] env[62066]: DEBUG nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 919.661452] env[62066]: INFO nova.compute.claims [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.665445] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.700483] env[62066]: INFO nova.compute.manager [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Took 0.56 seconds to detach 1 volumes for instance. [ 919.787741] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485486} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.788086] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 4983ece5-bb4b-43cb-a928-324b2f31030e/4983ece5-bb4b-43cb-a928-324b2f31030e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 919.788254] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.788513] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53997253-3890-4d36-af54-a75a11e61752 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.794549] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 919.794549] env[62066]: value = "task-1156452" [ 919.794549] env[62066]: _type = "Task" [ 919.794549] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.802826] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156452, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.923861] env[62066]: DEBUG nova.network.neutron [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updated VIF entry in instance network info cache for port 8cbada80-ab13-439c-9a90-eea92446a751. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 919.924384] env[62066]: DEBUG nova.network.neutron [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updating instance_info_cache with network_info: [{"id": "8cbada80-ab13-439c-9a90-eea92446a751", "address": "fa:16:3e:91:de:39", "network": {"id": "ba0b5add-84aa-416d-ac94-4ca1b048e258", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1395858552-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcef20dbc27940be90f355cc136bd436", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5291d0-ee0f-4d70-b2ae-ab6879a67b08", "external-id": "nsx-vlan-transportzone-597", "segmentation_id": 597, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cbada80-ab", "ovs_interfaceid": "8cbada80-ab13-439c-9a90-eea92446a751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.934679] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.970108] env[62066]: DEBUG nova.policy [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.977318] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156450, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.115273] env[62066]: DEBUG nova.objects.instance [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.212616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.304894] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156452, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066866} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.305330] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.306443] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d428b7fe-cc88-4592-9f17-4361d5556aa3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.336276] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 4983ece5-bb4b-43cb-a928-324b2f31030e/4983ece5-bb4b-43cb-a928-324b2f31030e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.339367] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47b01d92-d1d4-477b-8b65-5b1d4e1f443e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.358533] env[62066]: DEBUG nova.network.neutron [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.364597] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 920.364597] env[62066]: value = "task-1156453" [ 920.364597] env[62066]: _type = "Task" [ 920.364597] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.371258] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.427092] env[62066]: DEBUG oslo_concurrency.lockutils [req-5adf44e7-86b5-4ba9-8c8b-95ba225fcdab req-3a0d273f-cf44-4888-9934-fab6479c5e9e service nova] Releasing lock "refresh_cache-f82c8bf2-24f3-4732-a56d-1bddaef14158" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.477320] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156450, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.480563] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc80dde4-eef3-455c-8968-29fbc094dac3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.488509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9f19a8-782d-47f1-a5d1-94a461feda4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.522690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2581e990-0f30-4bf1-9254-8a78d0ec4a04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.530941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3c58b7-330f-4adc-a8d0-56899693cdcf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.547713] env[62066]: DEBUG nova.compute.provider_tree [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.618969] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.619190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.619355] env[62066]: DEBUG nova.network.neutron [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.619553] env[62066]: DEBUG nova.objects.instance [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'info_cache' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.862808] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.863159] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Instance network_info: |[{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.863723] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:ef:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aef860e5-c337-4d8f-b0ff-e3c92e1ff75f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.871230] env[62066]: DEBUG oslo.service.loopingcall [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.874817] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.875058] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7e160b1-c15f-4562-a119-a3d3eb6e927d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.894846] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156453, 'name': ReconfigVM_Task, 'duration_secs': 0.302286} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.899050] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 4983ece5-bb4b-43cb-a928-324b2f31030e/4983ece5-bb4b-43cb-a928-324b2f31030e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.899656] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.899656] env[62066]: value = "task-1156454" [ 920.899656] env[62066]: _type = "Task" [ 920.899656] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.899849] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-396eef6f-5783-4c9d-b276-918f0201606e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.914240] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 920.914240] env[62066]: value = "task-1156455" [ 920.914240] env[62066]: _type = "Task" [ 920.914240] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.914447] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156454, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.982226] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156450, 'name': Destroy_Task, 'duration_secs': 1.530979} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.983161] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Destroyed the VM [ 920.983161] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 920.983311] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-25d02ea3-901b-4942-872d-da729bd48e9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.997045] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 920.997045] env[62066]: value = "task-1156456" [ 920.997045] env[62066]: _type = "Task" [ 920.997045] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.005354] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156456, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.054783] env[62066]: DEBUG nova.scheduler.client.report [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.123126] env[62066]: DEBUG nova.objects.base [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Object Instance<31e50f97-f873-44ad-9923-67923cdb8d3a> lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 921.156967] env[62066]: DEBUG nova.compute.manager [req-3c7414e9-ad5b-4684-b82d-d7c18ccaed44 req-51ec9f9f-b656-4e25-aa32-0db6c709814f service nova] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Received event network-vif-deleted-4e3a7a96-3bfe-4fbe-9515-fd341f054700 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.412541] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156454, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.425592] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156455, 'name': Rename_Task, 'duration_secs': 0.197421} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.425898] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.426182] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5407df6b-3ee7-4262-837c-431be77cd8da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.433762] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 921.433762] env[62066]: value = "task-1156457" [ 921.433762] env[62066]: _type = "Task" [ 921.433762] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.445282] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.504768] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156456, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.511396] env[62066]: DEBUG nova.compute.manager [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Received event network-changed-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.511605] env[62066]: DEBUG nova.compute.manager [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Refreshing instance network info cache due to event network-changed-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.511830] env[62066]: DEBUG oslo_concurrency.lockutils [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] Acquiring lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.512870] env[62066]: DEBUG oslo_concurrency.lockutils [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] Acquired lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.513209] env[62066]: DEBUG nova.network.neutron [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Refreshing network info cache for port aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.560134] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.910s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.563335] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.046s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.565406] env[62066]: INFO nova.compute.claims [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.615511] env[62066]: INFO nova.network.neutron [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating port c9795c0a-c035-447a-9433-6cd7ff5ab2ef with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 921.913848] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156454, 'name': CreateVM_Task, 'duration_secs': 0.521355} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.915971] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.915971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.915971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.915971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.915971] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f934509-7f00-48b4-9e18-59c23fb248db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.920915] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 921.920915] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d78ff7-f114-eb4a-cadf-504010dbf94d" [ 921.920915] env[62066]: _type = "Task" [ 921.920915] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.930724] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d78ff7-f114-eb4a-cadf-504010dbf94d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.950319] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156457, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.011849] env[62066]: DEBUG oslo_vmware.api [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156456, 'name': RemoveSnapshot_Task, 'duration_secs': 0.854435} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.012189] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.122826] env[62066]: DEBUG nova.network.neutron [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.211332] env[62066]: DEBUG nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Successfully updated port: 5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.275547] env[62066]: DEBUG nova.network.neutron [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updated VIF entry in instance network info cache for port aef860e5-c337-4d8f-b0ff-e3c92e1ff75f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 922.276058] env[62066]: DEBUG nova.network.neutron [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.432603] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d78ff7-f114-eb4a-cadf-504010dbf94d, 'name': SearchDatastore_Task, 'duration_secs': 0.016417} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.432953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.433342] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.433648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.433836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.434027] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.434305] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51f88d75-8fa4-48d9-b258-3568da2c64f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.444368] env[62066]: DEBUG oslo_vmware.api [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156457, 'name': PowerOnVM_Task, 'duration_secs': 0.558095} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.444661] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.444874] env[62066]: INFO nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Took 8.18 seconds to spawn the instance on the hypervisor. [ 922.445057] env[62066]: DEBUG nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.445810] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2de760-486a-42db-8c75-27db02facb23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.449125] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.449276] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.450241] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d319b46d-aa41-4433-9d72-7a518a37a4a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.458846] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 922.458846] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ec32b8-7f5e-ff21-e817-804471eab00f" [ 922.458846] env[62066]: _type = "Task" [ 922.458846] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.466806] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ec32b8-7f5e-ff21-e817-804471eab00f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.523461] env[62066]: WARNING nova.compute.manager [None req-6e27c5da-14cd-4bac-82e8-067e3e672aa8 tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Image not found during snapshot: nova.exception.ImageNotFound: Image 84851c67-5ec7-44f1-906b-ef3db11ba04a could not be found. [ 922.626325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.716542] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.716748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.716998] env[62066]: DEBUG nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.766588] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183820b9-270b-4904-8744-8ec67e8c02fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.774900] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f0fd95-bbaf-41c0-ae5a-bd2a606f24bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.778961] env[62066]: DEBUG oslo_concurrency.lockutils [req-9d5bd761-342b-4114-8bd4-3e211018419d req-37fd8297-fef0-42bd-a81c-5aa449d4a804 service nova] Releasing lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.810418] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecd0417-e23c-43da-9c93-44d9ef427d04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.818466] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4e51a8-a923-461f-8aa0-4000996deb7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.832312] env[62066]: DEBUG nova.compute.provider_tree [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.967195] env[62066]: INFO nova.compute.manager [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Took 22.98 seconds to build instance. [ 922.972710] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ec32b8-7f5e-ff21-e817-804471eab00f, 'name': SearchDatastore_Task, 'duration_secs': 0.026738} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.973545] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a0f33af-708b-47fc-afa4-46fb583b2a25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.978920] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 922.978920] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5268ddfb-3a41-63c5-59b5-41661723ca98" [ 922.978920] env[62066]: _type = "Task" [ 922.978920] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.987740] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268ddfb-3a41-63c5-59b5-41661723ca98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.129497] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.129917] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd6f6f98-3152-4778-8770-8ef4c643a3b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.138347] env[62066]: DEBUG oslo_vmware.api [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 923.138347] env[62066]: value = "task-1156458" [ 923.138347] env[62066]: _type = "Task" [ 923.138347] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.152898] env[62066]: DEBUG oslo_vmware.api [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156458, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.247756] env[62066]: DEBUG nova.compute.manager [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.248036] env[62066]: DEBUG oslo_concurrency.lockutils [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.248269] env[62066]: DEBUG oslo_concurrency.lockutils [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.248660] env[62066]: DEBUG oslo_concurrency.lockutils [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.248873] env[62066]: DEBUG nova.compute.manager [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] No waiting events found dispatching network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.249072] env[62066]: WARNING nova.compute.manager [req-0d1a92b2-8b2a-4176-9b25-00d435d64a97 req-3e91fe16-0746-4618-94b8-0c3720f83270 service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received unexpected event network-vif-plugged-c9795c0a-c035-447a-9433-6cd7ff5ab2ef for instance with vm_state shelved_offloaded and task_state spawning. [ 923.276220] env[62066]: WARNING nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 923.276788] env[62066]: WARNING nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 923.321106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.321241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.322196] env[62066]: DEBUG nova.network.neutron [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.339018] env[62066]: DEBUG nova.scheduler.client.report [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.468887] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26101d31-bc58-49f3-8b0c-fd95a925ecfb tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.496s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.489957] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268ddfb-3a41-63c5-59b5-41661723ca98, 'name': SearchDatastore_Task, 'duration_secs': 0.03994} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.490407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.490544] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.490785] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f6efe6d-7593-4255-a2af-04fed6ee34be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.501256] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 923.501256] env[62066]: value = "task-1156459" [ 923.501256] env[62066]: _type = "Task" [ 923.501256] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.510379] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156459, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.650256] env[62066]: DEBUG oslo_vmware.api [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156458, 'name': PowerOnVM_Task, 'duration_secs': 0.401903} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.650674] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.650796] env[62066]: DEBUG nova.compute.manager [None req-85db3ed5-b2d6-4056-94bf-302fdaa81cb8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.651676] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0f0cd4-a6b4-4e4f-88df-7a8f744afb93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.740288] env[62066]: DEBUG nova.network.neutron [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38b4765c-0887-499c-b63b-895e1168aa5d", "address": "fa:16:3e:a7:45:38", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38b4765c-08", "ovs_interfaceid": "38b4765c-0887-499c-b63b-895e1168aa5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "address": "fa:16:3e:54:e6:d2", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fb737fe-ca", "ovs_interfaceid": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.820478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.820682] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.820848] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.821075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.821255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.823961] env[62066]: INFO nova.compute.manager [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Terminating instance [ 923.827831] env[62066]: DEBUG nova.compute.manager [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.828048] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.828918] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f639983-e573-48b9-88fb-2abd59c0d2f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.838470] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.838710] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3e679fa-5f74-4fd2-a4f2-77901423e21d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.841110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.841646] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.844555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.179s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.845229] env[62066]: DEBUG nova.objects.instance [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'resources' on Instance uuid 831c0bed-8a41-4672-aa08-d60b64f365b8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.853333] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 923.853333] env[62066]: value = "task-1156460" [ 923.853333] env[62066]: _type = "Task" [ 923.853333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.865256] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156460, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.917907] env[62066]: DEBUG nova.compute.manager [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-plugged-5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.918160] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.918436] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.918629] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.918877] env[62066]: DEBUG nova.compute.manager [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] No waiting events found dispatching network-vif-plugged-5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.919040] env[62066]: WARNING nova.compute.manager [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received unexpected event network-vif-plugged-5fb737fe-ca04-4ac5-851d-d2d9397ae02d for instance with vm_state active and task_state None. [ 923.919258] env[62066]: DEBUG nova.compute.manager [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-changed-5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.919456] env[62066]: DEBUG nova.compute.manager [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing instance network info cache due to event network-changed-5fb737fe-ca04-4ac5-851d-d2d9397ae02d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.919635] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.011409] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156459, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490732} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.011743] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.011978] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.012271] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a88bfa3d-9916-463a-aee2-9208bc05708b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.019849] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 924.019849] env[62066]: value = "task-1156461" [ 924.019849] env[62066]: _type = "Task" [ 924.019849] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.028827] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156461, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.069359] env[62066]: DEBUG nova.network.neutron [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.244223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.244874] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.245015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.245309] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.245502] env[62066]: DEBUG nova.network.neutron [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Refreshing network info cache for port 5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.247405] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f3c915-c074-48d0-af61-67484f120669 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.266652] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.266895] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.267072] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.267266] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.267420] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.267573] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.267785] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.267949] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.268153] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.268325] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.268501] env[62066]: DEBUG nova.virt.hardware [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.274814] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 924.275741] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94fe6d7f-6811-4e42-a80f-8099cbbd9f1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.294052] env[62066]: DEBUG oslo_vmware.api [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 924.294052] env[62066]: value = "task-1156462" [ 924.294052] env[62066]: _type = "Task" [ 924.294052] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.302358] env[62066]: DEBUG oslo_vmware.api [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156462, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.348650] env[62066]: DEBUG nova.compute.utils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 924.353591] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.353838] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 924.365899] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156460, 'name': PowerOffVM_Task, 'duration_secs': 0.274477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.366190] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.366362] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.366607] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00578a77-fddc-43d1-acc2-adfd73a1aa2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.403893] env[62066]: DEBUG nova.policy [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c41065776f39438a8f6d427c0f62b93f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c468d5ba348d437f97a74e0da70bb42e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.530639] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074178} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.530946] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.531773] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7d3d2f-e6a0-4b51-959f-8f1c52076018 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.553878] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.556585] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a9afe17-9be5-4edc-911b-ec48fae68b82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.571891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.577015] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 924.577015] env[62066]: value = "task-1156464" [ 924.577015] env[62066]: _type = "Task" [ 924.577015] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.591286] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156464, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.612219] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1224776d07d8a77c4031ec665a697a4d',container_format='bare',created_at=2024-10-07T00:58:35Z,direct_url=,disk_format='vmdk',id=c96c784e-9ccb-46d7-b666-d95d5fb00a8b,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-14742430-shelved',owner='d61d04b8fe994dcfaa361bb87eb587d6',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-07T00:58:52Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.612541] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.612737] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.612935] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.613097] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.613252] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.613470] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.613666] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.613855] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.614042] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.614228] env[62066]: DEBUG nova.virt.hardware [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.615127] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb03f6d4-51d3-4dcd-b8ac-5fa4767ce50c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.619074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8fb57e-6fb9-4f24-879d-39ded5ac8a3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.628968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f6c186-4c7e-4ab3-9a60-7870e675b8db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.633591] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1d4ac2-5a30-42b4-8472-b5b5f00e01c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.646758] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:3e:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e02e98f-44ce-42b7-a3ac-4034fae5d127', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9795c0a-c035-447a-9433-6cd7ff5ab2ef', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.654145] env[62066]: DEBUG oslo.service.loopingcall [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.680379] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.681195] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fefa7771-2966-4377-b53f-e3413f8b3468 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.696074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7ddf91-158d-4b7f-a821-4892697bad77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.704569] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a73e05-74fd-4599-8c5e-0d0dc1f5deea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.709901] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Successfully created port: f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.711705] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.711705] env[62066]: value = "task-1156465" [ 924.711705] env[62066]: _type = "Task" [ 924.711705] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.723370] env[62066]: DEBUG nova.compute.provider_tree [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.729225] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156465, 'name': CreateVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.779469] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.779693] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.779880] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleting the datastore file [datastore1] d8aa9d38-0a2c-4036-9574-a85d2ed83fb2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.780177] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50037838-9930-471f-b146-67ea359c3743 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.786091] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for the task: (returnval){ [ 924.786091] env[62066]: value = "task-1156466" [ 924.786091] env[62066]: _type = "Task" [ 924.786091] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.795394] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.803777] env[62066]: DEBUG oslo_vmware.api [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.806342] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "4983ece5-bb4b-43cb-a928-324b2f31030e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.806605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.806761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.806949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.807137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.811067] env[62066]: INFO nova.compute.manager [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Terminating instance [ 924.813021] env[62066]: DEBUG nova.compute.manager [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.813231] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.814022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e216485-2e60-4802-af80-7fada1fd4f05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.820339] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.820610] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-488d5cf7-83e5-476c-893a-587ad77b4cc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.826611] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 924.826611] env[62066]: value = "task-1156467" [ 924.826611] env[62066]: _type = "Task" [ 924.826611] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.834462] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.854281] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.087769] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156464, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.115992] env[62066]: DEBUG nova.network.neutron [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updated VIF entry in instance network info cache for port 5fb737fe-ca04-4ac5-851d-d2d9397ae02d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.116548] env[62066]: DEBUG nova.network.neutron [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38b4765c-0887-499c-b63b-895e1168aa5d", "address": "fa:16:3e:a7:45:38", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38b4765c-08", "ovs_interfaceid": "38b4765c-0887-499c-b63b-895e1168aa5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "address": "fa:16:3e:54:e6:d2", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fb737fe-ca", "ovs_interfaceid": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.224524] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156465, 'name': CreateVM_Task, 'duration_secs': 0.402371} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.224815] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.225879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.226212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.227410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.227410] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8312a529-fd5b-4793-9f45-bea9fb8c1f60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.231676] env[62066]: DEBUG nova.scheduler.client.report [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.236451] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 925.236451] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529e17ca-d4c7-ea92-63dd-22c5e277d5e4" [ 925.236451] env[62066]: _type = "Task" [ 925.236451] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.247633] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529e17ca-d4c7-ea92-63dd-22c5e277d5e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.297128] env[62066]: DEBUG oslo_vmware.api [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Task: {'id': task-1156466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253916} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.300429] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.300698] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.300881] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.301081] env[62066]: INFO nova.compute.manager [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Took 1.47 seconds to destroy the instance on the hypervisor. [ 925.301325] env[62066]: DEBUG oslo.service.loopingcall [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.301539] env[62066]: DEBUG nova.compute.manager [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.301629] env[62066]: DEBUG nova.network.neutron [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 925.309036] env[62066]: DEBUG oslo_vmware.api [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.336810] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156467, 'name': PowerOffVM_Task, 'duration_secs': 0.272312} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.338655] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.338655] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.338655] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8ad66bf-ba65-45e8-bfdb-5bbee3605d0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.401209] env[62066]: DEBUG nova.compute.manager [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.401209] env[62066]: DEBUG nova.compute.manager [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing instance network info cache due to event network-changed-c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.401209] env[62066]: DEBUG oslo_concurrency.lockutils [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.401209] env[62066]: DEBUG oslo_concurrency.lockutils [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.401481] env[62066]: DEBUG nova.network.neutron [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Refreshing network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.432028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.432028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.432028] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] 4983ece5-bb4b-43cb-a928-324b2f31030e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.432028] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-236ef9f2-05c9-4f11-b39f-97358a570fee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.438260] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 925.438260] env[62066]: value = "task-1156469" [ 925.438260] env[62066]: _type = "Task" [ 925.438260] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.446535] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.588172] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156464, 'name': ReconfigVM_Task, 'duration_secs': 0.583092} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.588484] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.589181] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-083490d1-1701-4c3c-85b6-1a292d51df35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.596115] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 925.596115] env[62066]: value = "task-1156470" [ 925.596115] env[62066]: _type = "Task" [ 925.596115] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.605961] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156470, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.619706] env[62066]: DEBUG oslo_concurrency.lockutils [req-a902438a-b65e-4519-b336-be19fc00fcd9 req-d00ac6a6-85e5-489a-9acc-231f6b9d21c2 service nova] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.737809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.740412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.530s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.740658] env[62066]: DEBUG nova.objects.instance [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'resources' on Instance uuid 7d7743d0-acf5-426e-b409-c00dca895048 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.752739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.752997] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Processing image c96c784e-9ccb-46d7-b666-d95d5fb00a8b {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.753260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.753411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.753595] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.753854] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33deb6ba-6fce-4bc4-a8a1-8145b9721ea8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.761805] env[62066]: INFO nova.scheduler.client.report [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted allocations for instance 831c0bed-8a41-4672-aa08-d60b64f365b8 [ 925.765851] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.765851] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.766748] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e671eeb-2083-4329-ab4a-40d74030b315 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.772431] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 925.772431] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52064b04-e7f0-c403-0c8b-7f322eb475aa" [ 925.772431] env[62066]: _type = "Task" [ 925.772431] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.780263] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52064b04-e7f0-c403-0c8b-7f322eb475aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.805544] env[62066]: DEBUG oslo_vmware.api [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156462, 'name': ReconfigVM_Task, 'duration_secs': 1.165201} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.806529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.806529] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 925.864117] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 925.892107] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.892378] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.892538] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.892811] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.892968] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.893151] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.893367] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.893531] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.893954] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.895179] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.895179] env[62066]: DEBUG nova.virt.hardware [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.895871] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e2251c-3abe-42f8-8bde-1399d7414d49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.906417] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3ab018-ba64-4af2-ac49-443c8d578fb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.948362] env[62066]: DEBUG oslo_vmware.api [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160053} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.948629] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.948823] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.949017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.949206] env[62066]: INFO nova.compute.manager [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 925.949458] env[62066]: DEBUG oslo.service.loopingcall [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.949893] env[62066]: DEBUG nova.compute.manager [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.949893] env[62066]: DEBUG nova.network.neutron [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.062216] env[62066]: DEBUG nova.network.neutron [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.101185] env[62066]: DEBUG nova.network.neutron [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updated VIF entry in instance network info cache for port c9795c0a-c035-447a-9433-6cd7ff5ab2ef. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.101708] env[62066]: DEBUG nova.network.neutron [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.108572] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156470, 'name': Rename_Task, 'duration_secs': 0.186861} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.108572] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.108572] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29798539-ceb6-46f6-820c-1811c41050d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.115407] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 926.115407] env[62066]: value = "task-1156471" [ 926.115407] env[62066]: _type = "Task" [ 926.115407] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.123404] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.252852] env[62066]: DEBUG nova.compute.manager [req-d70efa49-77a0-4c09-96bc-2e4325cda29c req-3783bde3-7927-4c0f-9cc1-6c2ef1186ae0 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Received event network-vif-deleted-30527f31-db37-41fe-931c-15eb75cc3794 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.253188] env[62066]: INFO nova.compute.manager [req-d70efa49-77a0-4c09-96bc-2e4325cda29c req-3783bde3-7927-4c0f-9cc1-6c2ef1186ae0 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Neutron deleted interface 30527f31-db37-41fe-931c-15eb75cc3794; detaching it from the instance and deleting it from the info cache [ 926.253506] env[62066]: DEBUG nova.network.neutron [req-d70efa49-77a0-4c09-96bc-2e4325cda29c req-3783bde3-7927-4c0f-9cc1-6c2ef1186ae0 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.274336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-40672802-eea8-44f5-8ab6-58c4295a779a tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "831c0bed-8a41-4672-aa08-d60b64f365b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.462s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.287623] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 926.288213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Fetch image to [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40/OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 926.288213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Downloading stream optimized image c96c784e-9ccb-46d7-b666-d95d5fb00a8b to [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40/OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40.vmdk on the data store datastore1 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 926.288365] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Downloading image file data c96c784e-9ccb-46d7-b666-d95d5fb00a8b to the ESX as VM named 'OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 926.317138] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea4e84d7-7d23-4743-8220-f71a92d8801e tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-5fb737fe-ca04-4ac5-851d-d2d9397ae02d" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.955s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.372814] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 926.372814] env[62066]: value = "resgroup-9" [ 926.372814] env[62066]: _type = "ResourcePool" [ 926.372814] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 926.372814] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-edcef4ef-ee7b-418a-a64b-343587cab827 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.395739] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lease: (returnval){ [ 926.395739] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 926.395739] env[62066]: _type = "HttpNfcLease" [ 926.395739] env[62066]: } obtained for vApp import into resource pool (val){ [ 926.395739] env[62066]: value = "resgroup-9" [ 926.395739] env[62066]: _type = "ResourcePool" [ 926.395739] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 926.396186] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the lease: (returnval){ [ 926.396186] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 926.396186] env[62066]: _type = "HttpNfcLease" [ 926.396186] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 926.405612] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.405612] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 926.405612] env[62066]: _type = "HttpNfcLease" [ 926.405612] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 926.418934] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Successfully updated port: f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.526829] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33078401-95f9-44e9-9c08-a1ca3558c5df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.536264] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1211b822-67ff-4883-ad16-55d931e7153e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.568231] env[62066]: INFO nova.compute.manager [-] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Took 1.27 seconds to deallocate network for instance. [ 926.571182] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141f441f-13ba-43cb-ab5d-3ef2770c160b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.581425] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caea3fc-3584-40ad-b9af-6dfcd406e2da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.596180] env[62066]: DEBUG nova.compute.provider_tree [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.603895] env[62066]: DEBUG oslo_concurrency.lockutils [req-09ce5b21-e89e-4417-a656-01dc72e72ab5 req-a957b254-ac1d-42ef-bc91-8226335a90fc service nova] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.625537] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156471, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.704784] env[62066]: DEBUG nova.network.neutron [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.756215] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f05b39b5-4cd8-436a-809a-1a461083af8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.765426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aaa7fdc-320a-406b-aa98-22057ee45cb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.796639] env[62066]: DEBUG nova.compute.manager [req-d70efa49-77a0-4c09-96bc-2e4325cda29c req-3783bde3-7927-4c0f-9cc1-6c2ef1186ae0 service nova] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Detach interface failed, port_id=30527f31-db37-41fe-931c-15eb75cc3794, reason: Instance 4983ece5-bb4b-43cb-a928-324b2f31030e could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 926.906712] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.906712] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 926.906712] env[62066]: _type = "HttpNfcLease" [ 926.906712] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 926.927519] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.927685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.927870] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.078959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.099153] env[62066]: DEBUG nova.scheduler.client.report [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.132630] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156471, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.209624] env[62066]: INFO nova.compute.manager [-] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Took 1.26 seconds to deallocate network for instance. [ 927.216599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-38b4765c-0887-499c-b63b-895e1168aa5d" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.217053] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-38b4765c-0887-499c-b63b-895e1168aa5d" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.404529] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.404529] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 927.404529] env[62066]: _type = "HttpNfcLease" [ 927.404529] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 927.431121] env[62066]: DEBUG nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Received event network-vif-deleted-d1a44472-08ab-45ac-94eb-3a8b9bb7ee35 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.431264] env[62066]: DEBUG nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Received event network-vif-plugged-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.431470] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.431610] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.431786] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.431954] env[62066]: DEBUG nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] No waiting events found dispatching network-vif-plugged-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.432140] env[62066]: WARNING nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Received unexpected event network-vif-plugged-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef for instance with vm_state building and task_state spawning. [ 927.432310] env[62066]: DEBUG nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Received event network-changed-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.432470] env[62066]: DEBUG nova.compute.manager [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Refreshing instance network info cache due to event network-changed-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.432685] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.463115] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.599487] env[62066]: DEBUG nova.network.neutron [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.604720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.606899] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.528s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.607538] env[62066]: DEBUG nova.objects.instance [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lazy-loading 'resources' on Instance uuid d8aa9d38-0a2c-4036-9574-a85d2ed83fb2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.627989] env[62066]: DEBUG oslo_vmware.api [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156471, 'name': PowerOnVM_Task, 'duration_secs': 1.244187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.628486] env[62066]: INFO nova.scheduler.client.report [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 7d7743d0-acf5-426e-b409-c00dca895048 [ 927.629456] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.629626] env[62066]: INFO nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Took 10.88 seconds to spawn the instance on the hypervisor. [ 927.629813] env[62066]: DEBUG nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.633590] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8a96cf-6943-4be3-a474-6792d890c047 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.720033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.720582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.720828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.721684] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c065e6-ed5b-4a90-b225-fa8c17209336 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.742381] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794e6c5f-50c1-4690-bc00-b96281ea04b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.770619] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 927.770932] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a808c743-3437-4138-a520-f54e1eb980bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.789286] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 927.789286] env[62066]: value = "task-1156474" [ 927.789286] env[62066]: _type = "Task" [ 927.789286] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.797211] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.905101] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.905101] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 927.905101] env[62066]: _type = "HttpNfcLease" [ 927.905101] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 927.905415] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 927.905415] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52244c0b-4fb3-cfcc-40a2-95a44ec06312" [ 927.905415] env[62066]: _type = "HttpNfcLease" [ 927.905415] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 927.906142] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee9876f-90b1-4d72-86b7-b99751cce2d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.913126] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 927.913311] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 927.978574] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3420d629-ac11-49ed-905c-e28944fc234f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.102746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.103180] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Instance network_info: |[{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.103578] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.103902] env[62066]: DEBUG nova.network.neutron [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Refreshing network info cache for port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 928.105620] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:a3:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.113087] env[62066]: DEBUG oslo.service.loopingcall [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.116785] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.119862] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2342351f-9108-4d03-8ee6-b712321e8694 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.142328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62376cf6-cbbb-4722-b2ea-d03219af20ac tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7d7743d0-acf5-426e-b409-c00dca895048" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.927s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.148839] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.148839] env[62066]: value = "task-1156475" [ 928.148839] env[62066]: _type = "Task" [ 928.148839] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.155438] env[62066]: INFO nova.compute.manager [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Took 23.09 seconds to build instance. [ 928.167977] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156475, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.304976] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.346115] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e356496b-c0cb-46be-b1dc-48605028dd9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.359282] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def889d7-e807-4a74-8fc3-f59594a5fea8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.408356] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733bd6d1-8158-464f-a3bf-37bb78d787e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.418689] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3094b0ec-0e8f-4c93-9d36-dd677d5b3680 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.436127] env[62066]: DEBUG nova.compute.provider_tree [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.664673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30d43495-2b8c-4dd6-8fc4-3776fcd673de tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.608s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.664673] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156475, 'name': CreateVM_Task, 'duration_secs': 0.343506} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.666632] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.667486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.667670] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.668152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.668561] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae725688-3f30-49e8-9952-04ba74b0e9b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.674384] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 928.674384] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5278fcf0-aef9-293a-bb11-aba1bff0ecdd" [ 928.674384] env[62066]: _type = "Task" [ 928.674384] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.685867] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5278fcf0-aef9-293a-bb11-aba1bff0ecdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.801545] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.845855] env[62066]: DEBUG nova.network.neutron [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updated VIF entry in instance network info cache for port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 928.846319] env[62066]: DEBUG nova.network.neutron [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.942566] env[62066]: DEBUG nova.scheduler.client.report [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.092438] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 929.092579] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 929.093531] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a89611d-31d9-43fa-9d4d-941f3a5457e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.100188] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 929.100601] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 929.100601] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b297be5d-6feb-4f0e-9659-dcefa91fc1b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.185954] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5278fcf0-aef9-293a-bb11-aba1bff0ecdd, 'name': SearchDatastore_Task, 'duration_secs': 0.010608} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.186365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.186642] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.186922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.187126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.187356] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.187649] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-408b8963-6186-471e-bd5c-793ac9f0bfd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.197323] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.197530] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.198329] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b4467bd-d31c-490e-919e-5f6c7d25a084 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.206981] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 929.206981] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cda4ff-f461-219b-a837-aa297f7ae40b" [ 929.206981] env[62066]: _type = "Task" [ 929.206981] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.218984] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cda4ff-f461-219b-a837-aa297f7ae40b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.299464] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.304720] env[62066]: DEBUG oslo_vmware.rw_handles [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527ed8c8-85c7-5e63-57d1-e37ad9ff31f5/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 929.305070] env[62066]: INFO nova.virt.vmwareapi.images [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Downloaded image file data c96c784e-9ccb-46d7-b666-d95d5fb00a8b [ 929.305697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abff22d7-3511-435f-acec-e896cc53de00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.321890] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fc4c63e-c57d-48cd-8251-3fa74279dba6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.342592] env[62066]: INFO nova.virt.vmwareapi.images [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] The imported VM was unregistered [ 929.345408] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 929.345925] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.346409] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a615f9a0-2871-4c0b-ba84-86982c822433 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.350315] env[62066]: DEBUG oslo_concurrency.lockutils [req-75720da7-9f99-4bc8-913a-7bf476a5abf8 req-a8622bb5-47c0-4f13-b5e0-4025ecbbfaf1 service nova] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.361020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Created directory with path [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.361020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40/OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40.vmdk to [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 929.361020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-b09ff507-1d58-4557-b646-2a6f6bcba929 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.368021] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 929.368021] env[62066]: value = "task-1156477" [ 929.368021] env[62066]: _type = "Task" [ 929.368021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.376613] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.448300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.450500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.731s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.450850] env[62066]: DEBUG nova.objects.instance [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'resources' on Instance uuid 4983ece5-bb4b-43cb-a928-324b2f31030e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.465122] env[62066]: INFO nova.scheduler.client.report [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Deleted allocations for instance d8aa9d38-0a2c-4036-9574-a85d2ed83fb2 [ 929.718959] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cda4ff-f461-219b-a837-aa297f7ae40b, 'name': SearchDatastore_Task, 'duration_secs': 0.011465} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.719823] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59eed042-f157-42bd-8df5-ef15932ff8e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.725385] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 929.725385] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffcd9f-3942-dca9-65a6-cdfda9b26949" [ 929.725385] env[62066]: _type = "Task" [ 929.725385] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.733714] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffcd9f-3942-dca9-65a6-cdfda9b26949, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.801941] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.878969] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.939614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.939614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.972954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-eb2e3a6f-892d-499a-8957-6a8bf09efe2c tempest-ImagesTestJSON-1773160046 tempest-ImagesTestJSON-1773160046-project-member] Lock "d8aa9d38-0a2c-4036-9574-a85d2ed83fb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.152s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.171123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d641752-95e6-44d4-b8b8-0c5131b4e819 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.180335] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f80df1-edff-4126-8c83-c0242073940c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.216958] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b469ba-2598-4399-af89-52a6c1eafd5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.225595] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54833ff0-4c76-4419-947f-5fce831557ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.243842] env[62066]: DEBUG nova.compute.provider_tree [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.249503] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffcd9f-3942-dca9-65a6-cdfda9b26949, 'name': SearchDatastore_Task, 'duration_secs': 0.083294} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.251525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.251525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.251525] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62179894-5a6d-47da-a86c-d34041b64679 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.259958] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 930.259958] env[62066]: value = "task-1156478" [ 930.259958] env[62066]: _type = "Task" [ 930.259958] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.271636] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.288032] env[62066]: DEBUG nova.compute.manager [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 930.301907] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.378837] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.443602] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.751808] env[62066]: DEBUG nova.scheduler.client.report [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.775102] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.805704] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.811690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.879149] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.974038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.262690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.266033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.454s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.277876] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.295985] env[62066]: INFO nova.scheduler.client.report [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 4983ece5-bb4b-43cb-a928-324b2f31030e [ 931.312996] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.381178] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.776272] env[62066]: INFO nova.compute.claims [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.785309] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156478, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.813486] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.815719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f48a75d9-fbdb-4739-bd05-5bbb90c73174 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "4983ece5-bb4b-43cb-a928-324b2f31030e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.880797] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156477, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.44448} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.881304] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40/OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40.vmdk to [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk. [ 931.881737] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Cleaning up location [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 931.882034] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_53602879-7fe4-460b-ba55-0ec2e14bad40 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.882446] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5b1326d-0a1d-4f26-b944-3818f42a857a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.891835] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 931.891835] env[62066]: value = "task-1156479" [ 931.891835] env[62066]: _type = "Task" [ 931.891835] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.902528] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.131507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "a8dd7483-0588-4f60-9504-20de799e69f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.131791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.131991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.132174] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.132347] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.134377] env[62066]: INFO nova.compute.manager [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Terminating instance [ 932.136342] env[62066]: DEBUG nova.compute.manager [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.136446] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.138030] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfeec31b-3d1a-46ab-83da-4fba0be496a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.146334] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.146580] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-988b238f-002e-4555-8591-5011a3ac8f2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.152403] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 932.152403] env[62066]: value = "task-1156480" [ 932.152403] env[62066]: _type = "Task" [ 932.152403] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.160929] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.281584] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156478, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.724307} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.282077] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.282546] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.282935] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-593453d5-132c-4038-b19d-318ddc98337b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.289138] env[62066]: INFO nova.compute.resource_tracker [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating resource usage from migration 7806705f-7018-4c75-92eb-f4d8a764f5ed [ 932.293710] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 932.293710] env[62066]: value = "task-1156481" [ 932.293710] env[62066]: _type = "Task" [ 932.293710] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.312882] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156481, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.316089] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.399856] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095783} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.402267] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.402451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.402719] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk to [datastore1] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.403172] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ca7fee9-2107-4596-bd28-6896e8ae972d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.410285] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 932.410285] env[62066]: value = "task-1156482" [ 932.410285] env[62066]: _type = "Task" [ 932.410285] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.421298] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.486548] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9a4ce5-f03e-453c-b1a6-f5121a857639 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.494079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4a32a0-5ce2-4764-9945-3fe498830476 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.524505] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a016c484-b390-4c02-8c9a-fc3d0a8bfffe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.531845] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fd1376-9f92-4d5d-9747-77d69220a02d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.546680] env[62066]: DEBUG nova.compute.provider_tree [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.627502] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "b2afea5a-2390-40b0-b860-c50f71ebac94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.627824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.628057] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.628261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.628430] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.630821] env[62066]: INFO nova.compute.manager [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Terminating instance [ 932.633393] env[62066]: DEBUG nova.compute.manager [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.633628] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.635254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d727bf-e417-44fd-af4c-f7b52ffa4375 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.642643] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.642964] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04a23019-dc48-4ab0-b258-0f09fae6cdde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.649405] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 932.649405] env[62066]: value = "task-1156483" [ 932.649405] env[62066]: _type = "Task" [ 932.649405] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.659428] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.664646] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156480, 'name': PowerOffVM_Task, 'duration_secs': 0.451988} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.664960] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.665169] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.665448] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63d92bc1-007f-4f8a-8b72-9db3d9496641 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.724160] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.724419] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.724605] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleting the datastore file [datastore2] a8dd7483-0588-4f60-9504-20de799e69f1 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.724920] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a000f979-03ec-4797-a437-5cf75a2eff59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.732147] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 932.732147] env[62066]: value = "task-1156485" [ 932.732147] env[62066]: _type = "Task" [ 932.732147] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.744693] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.806363] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156481, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06681} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.810575] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.811655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0358c7a1-9a78-49bc-80fd-410046aeb45e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.821431] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.841977] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.842425] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d146207e-60e4-4d62-99d6-619aa9ca98f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.863910] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 932.863910] env[62066]: value = "task-1156486" [ 932.863910] env[62066]: _type = "Task" [ 932.863910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.873038] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156486, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.921303] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.052281] env[62066]: DEBUG nova.scheduler.client.report [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.170193] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156483, 'name': PowerOffVM_Task, 'duration_secs': 0.310036} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.170193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.170193] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.170193] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afeda1b8-3f0e-4092-8a50-5e032909a2b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.246798] env[62066]: DEBUG oslo_vmware.api [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318338} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.246798] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.246798] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.246798] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.246798] env[62066]: INFO nova.compute.manager [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 933.246798] env[62066]: DEBUG oslo.service.loopingcall [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.246798] env[62066]: DEBUG nova.compute.manager [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.246798] env[62066]: DEBUG nova.network.neutron [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 933.318339] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.319811] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.320226] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] b2afea5a-2390-40b0-b860-c50f71ebac94 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.322370] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc62ccb2-c751-460b-b7b4-11c1a08193aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329883] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.336931] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 933.336931] env[62066]: value = "task-1156488" [ 933.336931] env[62066]: _type = "Task" [ 933.336931] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.347845] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.375247] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156486, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.422335] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.560352] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.294s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.560571] env[62066]: INFO nova.compute.manager [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Migrating [ 933.569713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.595s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.570628] env[62066]: INFO nova.compute.claims [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.820259] env[62066]: DEBUG oslo_vmware.api [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156474, 'name': ReconfigVM_Task, 'duration_secs': 5.811014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.820530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.820783] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 933.847394] env[62066]: DEBUG oslo_vmware.api [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244033} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.848762] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.848880] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.849190] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.849449] env[62066]: INFO nova.compute.manager [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Took 1.22 seconds to destroy the instance on the hypervisor. [ 933.849717] env[62066]: DEBUG oslo.service.loopingcall [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.850440] env[62066]: DEBUG nova.compute.manager [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.850544] env[62066]: DEBUG nova.network.neutron [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 933.872038] env[62066]: DEBUG nova.compute.manager [req-aa48b598-89e2-4d27-902b-131746940739 req-e4dcbaaf-bffa-4089-98f2-05d5f6fac2cf service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Received event network-vif-deleted-983e5ae8-8e7c-41f2-9304-39f8dc1b8347 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.872098] env[62066]: INFO nova.compute.manager [req-aa48b598-89e2-4d27-902b-131746940739 req-e4dcbaaf-bffa-4089-98f2-05d5f6fac2cf service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Neutron deleted interface 983e5ae8-8e7c-41f2-9304-39f8dc1b8347; detaching it from the instance and deleting it from the info cache [ 933.872261] env[62066]: DEBUG nova.network.neutron [req-aa48b598-89e2-4d27-902b-131746940739 req-e4dcbaaf-bffa-4089-98f2-05d5f6fac2cf service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.881346] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156486, 'name': ReconfigVM_Task, 'duration_secs': 0.593949} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.881639] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.882297] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec8d9a84-3b58-4b5b-9a02-a666e2acd6db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.889778] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 933.889778] env[62066]: value = "task-1156489" [ 933.889778] env[62066]: _type = "Task" [ 933.889778] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.903303] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156489, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.922574] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.085307] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.085546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.085711] env[62066]: DEBUG nova.network.neutron [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.140321] env[62066]: DEBUG nova.compute.manager [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-deleted-38b4765c-0887-499c-b63b-895e1168aa5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.140576] env[62066]: INFO nova.compute.manager [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Neutron deleted interface 38b4765c-0887-499c-b63b-895e1168aa5d; detaching it from the instance and deleting it from the info cache [ 934.145764] env[62066]: DEBUG nova.network.neutron [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "address": "fa:16:3e:54:e6:d2", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fb737fe-ca", "ovs_interfaceid": "5fb737fe-ca04-4ac5-851d-d2d9397ae02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.338493] env[62066]: DEBUG nova.network.neutron [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.382788] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc6b4afb-9bd0-4c7f-889e-26dc99b8f728 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.394195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c830c2fb-e733-476c-9ce4-991263448e3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.412726] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156489, 'name': Rename_Task, 'duration_secs': 0.260204} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.413099] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.413412] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4f63b4d-4d4b-41b2-bcd3-6822f6aa92ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.435785] env[62066]: DEBUG nova.compute.manager [req-aa48b598-89e2-4d27-902b-131746940739 req-e4dcbaaf-bffa-4089-98f2-05d5f6fac2cf service nova] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Detach interface failed, port_id=983e5ae8-8e7c-41f2-9304-39f8dc1b8347, reason: Instance a8dd7483-0588-4f60-9504-20de799e69f1 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 934.442152] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.443672] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 934.443672] env[62066]: value = "task-1156490" [ 934.443672] env[62066]: _type = "Task" [ 934.443672] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.453357] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.649902] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.650500] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] Acquired lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.651460] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06a917f-55ef-47a8-bfd3-fe1173273cc3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.675203] env[62066]: DEBUG oslo_concurrency.lockutils [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] Releasing lock "f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.675737] env[62066]: WARNING nova.compute.manager [req-7f50ba4b-e6a2-4ec1-8f82-ca769338c377 req-47059c57-b42e-4b13-88f8-a37fee0a385d service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Detach interface failed, port_id=38b4765c-0887-499c-b63b-895e1168aa5d, reason: No device with interface-id 38b4765c-0887-499c-b63b-895e1168aa5d exists on VM: nova.exception.NotFound: No device with interface-id 38b4765c-0887-499c-b63b-895e1168aa5d exists on VM [ 934.769028] env[62066]: DEBUG nova.network.neutron [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.843906] env[62066]: INFO nova.compute.manager [-] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Took 1.60 seconds to deallocate network for instance. [ 934.877591] env[62066]: DEBUG nova.network.neutron [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.887940] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b689cbd3-a499-48d0-871f-f8fb43170d39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.899935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75a2a6b-3d28-4b2e-b530-21c1299e4b9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.936585] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cfc735-4625-4d05-8076-b3fd027fee12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.946493] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156482, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.396669} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.949742] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c96c784e-9ccb-46d7-b666-d95d5fb00a8b/c96c784e-9ccb-46d7-b666-d95d5fb00a8b.vmdk to [datastore1] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.951417] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea71b2c-9e59-4941-9b16-5f89081c1106 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.955347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5456562f-54b2-4aa3-b468-c259b364b114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.962971] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156490, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.988449] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.988983] env[62066]: DEBUG nova.compute.provider_tree [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.990219] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72d302e4-c00f-469d-9f35-1332dc952db0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.010230] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 935.010230] env[62066]: value = "task-1156491" [ 935.010230] env[62066]: _type = "Task" [ 935.010230] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.020602] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.117888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.118158] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.118360] env[62066]: DEBUG nova.network.neutron [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.273088] env[62066]: INFO nova.compute.manager [-] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Took 1.42 seconds to deallocate network for instance. [ 935.351143] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.384664] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.386852] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.387894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.388300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.388652] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.388974] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.392972] env[62066]: INFO nova.compute.manager [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Terminating instance [ 935.395979] env[62066]: DEBUG nova.compute.manager [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.396146] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.397404] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154dffcb-ea1c-4e3d-b64b-3e5791cacc05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.410204] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.412304] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7279b0b-2f27-45c1-91f1-e7ce1081b48b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.425279] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 935.425279] env[62066]: value = "task-1156492" [ 935.425279] env[62066]: _type = "Task" [ 935.425279] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.438679] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156492, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.464030] env[62066]: DEBUG oslo_vmware.api [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156490, 'name': PowerOnVM_Task, 'duration_secs': 1.013108} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.464030] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.464030] env[62066]: INFO nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Took 9.60 seconds to spawn the instance on the hypervisor. [ 935.464030] env[62066]: DEBUG nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.465870] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89991a11-f535-478f-8eff-0de97918a53e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.506943] env[62066]: DEBUG nova.scheduler.client.report [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.527099] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156491, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.781166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.891725] env[62066]: INFO nova.network.neutron [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Port 5fb737fe-ca04-4ac5-851d-d2d9397ae02d from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 935.893147] env[62066]: DEBUG nova.network.neutron [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [{"id": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "address": "fa:16:3e:ce:60:da", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d8007bd-9e", "ovs_interfaceid": "0d8007bd-9e20-4780-a21e-a22c8c7dac13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.909704] env[62066]: DEBUG nova.compute.manager [req-199b19c8-0dde-4f9b-a92d-0aa9136adbe2 req-33571a21-cf09-4de6-96c1-206aec4c8c81 service nova] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Received event network-vif-deleted-cc490fbd-bf0f-41fd-a20a-8438810bab18 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.941096] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156492, 'name': PowerOffVM_Task, 'duration_secs': 0.486757} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.941096] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.941096] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.941096] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28a8ce47-be7d-406f-8650-42f875d5373b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.992923] env[62066]: INFO nova.compute.manager [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Took 21.49 seconds to build instance. [ 936.020021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.020021] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 936.021982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.671s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.022458] env[62066]: DEBUG nova.objects.instance [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'resources' on Instance uuid a8dd7483-0588-4f60-9504-20de799e69f1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.030357] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156491, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.062975] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.062975] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.062975] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleting the datastore file [datastore2] f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.062975] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-338b27c9-afb3-40dc-b031-4d61817f74b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.072697] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 936.072697] env[62066]: value = "task-1156494" [ 936.072697] env[62066]: _type = "Task" [ 936.072697] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.082968] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156494, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.234805] env[62066]: DEBUG nova.compute.manager [req-07453f2d-a7bf-46cc-a81a-22fef70221a6 req-719b60e7-d0c4-4302-9953-a9dfadcfd377 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-deleted-5fb737fe-ca04-4ac5-851d-d2d9397ae02d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.399546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-f24f8067-07b2-4941-8464-c30b58983ec1" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.496793] env[62066]: DEBUG oslo_concurrency.lockutils [None req-79a16385-13ac-455f-b4fc-0ad757b70ca3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.003s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.529948] env[62066]: DEBUG nova.compute.utils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 936.532459] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156491, 'name': ReconfigVM_Task, 'duration_secs': 1.170863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.533910] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.533910] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 936.537855] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76/0ba970e7-6b21-441a-81f7-2b4e7dfd4d76.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.542090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-683f0f97-e0f7-457e-a45e-a39d32db2dbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.550659] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 936.550659] env[62066]: value = "task-1156495" [ 936.550659] env[62066]: _type = "Task" [ 936.550659] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.562699] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156495, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.585198] env[62066]: DEBUG oslo_vmware.api [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156494, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.508917} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.587103] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.587408] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.587643] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.587868] env[62066]: INFO nova.compute.manager [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Took 1.19 seconds to destroy the instance on the hypervisor. [ 936.588177] env[62066]: DEBUG oslo.service.loopingcall [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.589626] env[62066]: DEBUG nova.compute.manager [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 936.589626] env[62066]: DEBUG nova.network.neutron [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 936.611194] env[62066]: DEBUG nova.policy [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.656209] env[62066]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 5fb737fe-ca04-4ac5-851d-d2d9397ae02d could not be found.", "detail": ""}} {{(pid=62066) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 936.656499] env[62066]: DEBUG nova.network.neutron [-] Unable to show port 5fb737fe-ca04-4ac5-851d-d2d9397ae02d as it no longer exists. {{(pid=62066) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 936.846605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "16390d89-f2d1-4d6c-be62-b85e61906865" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.846972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.849842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9b59da-80cf-43cb-9ede-b80dc5fe0864 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.866435] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5f1af1-9035-4902-8566-ebbe5548cd49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.902140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7235a5-26b8-42e8-bf69-ddc945b62d4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.906252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf0ff3e8-71e8-4b32-90db-269c1fbf3361 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-f24f8067-07b2-4941-8464-c30b58983ec1-38b4765c-0887-499c-b63b-895e1168aa5d" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.689s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.918042] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cae9e0-0fba-490d-9f59-51abad7b5aa9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.922199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b915f4cf-e41a-40a2-975d-de4ed5d6a258 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.953926] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.960499] env[62066]: DEBUG nova.compute.provider_tree [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.034331] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 937.063033] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156495, 'name': Rename_Task, 'duration_secs': 0.30296} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.063033] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.063033] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13f83031-65e1-4ce4-956f-6629dad0cb3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.069172] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 937.069172] env[62066]: value = "task-1156496" [ 937.069172] env[62066]: _type = "Task" [ 937.069172] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.076530] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Successfully created port: 14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.085872] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156496, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.358278] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.461124] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.461973] env[62066]: DEBUG nova.scheduler.client.report [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.465234] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fba70b4-f49f-4a60-8f9d-fe9466c2a7bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.473631] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 937.473631] env[62066]: value = "task-1156497" [ 937.473631] env[62066]: _type = "Task" [ 937.473631] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.484935] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.584975] env[62066]: DEBUG oslo_vmware.api [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156496, 'name': PowerOnVM_Task, 'duration_secs': 0.462532} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.584975] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.672679] env[62066]: DEBUG nova.network.neutron [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.710298] env[62066]: DEBUG nova.compute.manager [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.711098] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1512f033-c824-446a-a053-1f8159562e02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.889253] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.939877] env[62066]: DEBUG nova.compute.manager [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Received event network-changed-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.940221] env[62066]: DEBUG nova.compute.manager [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Refreshing instance network info cache due to event network-changed-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.940327] env[62066]: DEBUG oslo_concurrency.lockutils [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.940457] env[62066]: DEBUG oslo_concurrency.lockutils [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.940622] env[62066]: DEBUG nova.network.neutron [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Refreshing network info cache for port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.969077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.971515] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.191s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.971794] env[62066]: DEBUG nova.objects.instance [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'resources' on Instance uuid b2afea5a-2390-40b0-b860-c50f71ebac94 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.985313] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156497, 'name': PowerOffVM_Task, 'duration_secs': 0.319501} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.985591] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.985799] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.989733] env[62066]: INFO nova.scheduler.client.report [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted allocations for instance a8dd7483-0588-4f60-9504-20de799e69f1 [ 938.044622] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 938.067337] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.067629] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.067795] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.068016] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.068162] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.068354] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.068587] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.068754] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.068929] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.069145] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.069351] env[62066]: DEBUG nova.virt.hardware [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.070342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6099d2-dd8d-4a36-9b3f-802e88b1d66a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.080486] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374fe852-a6ca-4693-abd1-0ce1972feaba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.177200] env[62066]: INFO nova.compute.manager [-] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Took 1.59 seconds to deallocate network for instance. [ 938.230242] env[62066]: DEBUG oslo_concurrency.lockutils [None req-83046bc2-1974-4697-a8f2-b8d192f5c259 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.773s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.262623] env[62066]: DEBUG nova.compute.manager [req-f72d461a-5c66-448b-9ec7-8922fd3626d4 req-e0a59f7a-d651-4310-be60-6bc51af51c38 service nova] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Received event network-vif-deleted-0d8007bd-9e20-4780-a21e-a22c8c7dac13 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.500953] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.501462] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.511748] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f33ff5b-2f4b-426c-a884-b616bc70651c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.522942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-217e4ba1-d132-4068-81c5-6c360c3aefe2 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "a8dd7483-0588-4f60-9504-20de799e69f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.391s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.530140] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 938.530140] env[62066]: value = "task-1156498" [ 938.530140] env[62066]: _type = "Task" [ 938.530140] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.550148] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156498, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.663304] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Successfully updated port: 14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.685384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.743087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9058cf00-644c-4854-85b8-339b71f94db6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.754889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8a6d46-83c1-46ec-b511-647cfbdeea3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.791300] env[62066]: DEBUG nova.network.neutron [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updated VIF entry in instance network info cache for port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.791688] env[62066]: DEBUG nova.network.neutron [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.793452] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f694ba-5871-4002-a1a7-0e715a28a708 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.801966] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1a12e1-e18e-4046-99e8-dc581b412d4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.816715] env[62066]: DEBUG nova.compute.provider_tree [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.040177] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156498, 'name': ReconfigVM_Task, 'duration_secs': 0.262581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.040504] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.173149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.173380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.173632] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.251418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "d50f0505-66f4-412f-9744-25e0ea96277a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.251820] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.252076] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.252284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.252458] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.254620] env[62066]: INFO nova.compute.manager [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Terminating instance [ 939.256527] env[62066]: DEBUG nova.compute.manager [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.256724] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.257560] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe324cd-c2b7-43ce-863a-c722389326a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.265603] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.265829] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3827e948-0a19-4700-9763-42b8ab309b66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.272398] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 939.272398] env[62066]: value = "task-1156499" [ 939.272398] env[62066]: _type = "Task" [ 939.272398] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.281150] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.297047] env[62066]: DEBUG oslo_concurrency.lockutils [req-44b922e0-2885-4ed9-97dd-054c7415be44 req-09e616ee-9706-4c3e-a529-0c00f74461d4 service nova] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.319457] env[62066]: DEBUG nova.scheduler.client.report [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.547837] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.547837] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.547837] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.548483] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.548902] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.549275] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.551029] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.551029] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.551029] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.551029] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.551029] env[62066]: DEBUG nova.virt.hardware [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.555964] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 939.556283] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64414cd3-a78e-48e0-97f2-b3d19e80a37d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.575436] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 939.575436] env[62066]: value = "task-1156500" [ 939.575436] env[62066]: _type = "Task" [ 939.575436] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.583566] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156500, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.680815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820e0507-1736-4634-9f29-d2cafe0ab5b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.687699] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 939.687979] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-170f102f-e73c-46c2-a709-af0e0f667495 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.694250] env[62066]: DEBUG oslo_vmware.api [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 939.694250] env[62066]: value = "task-1156501" [ 939.694250] env[62066]: _type = "Task" [ 939.694250] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.702397] env[62066]: DEBUG oslo_vmware.api [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156501, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.708690] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.785206] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156499, 'name': PowerOffVM_Task, 'duration_secs': 0.256689} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.785573] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.785794] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.786084] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2eb8872d-2f11-490e-ae57-5920df34a3ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.825165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.827331] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.828894] env[62066]: INFO nova.compute.claims [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.850122] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.850399] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.850625] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleting the datastore file [datastore2] d50f0505-66f4-412f-9744-25e0ea96277a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.851601] env[62066]: INFO nova.scheduler.client.report [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance b2afea5a-2390-40b0-b860-c50f71ebac94 [ 939.852602] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d218d13-df01-46c5-b16b-a0016fe63b1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.861402] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for the task: (returnval){ [ 939.861402] env[62066]: value = "task-1156503" [ 939.861402] env[62066]: _type = "Task" [ 939.861402] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.869704] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.874953] env[62066]: DEBUG nova.network.neutron [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updating instance_info_cache with network_info: [{"id": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "address": "fa:16:3e:6a:8b:8e", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14b4402d-b3", "ovs_interfaceid": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.970632] env[62066]: DEBUG nova.compute.manager [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received event network-vif-plugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.970888] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.971162] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.971406] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.971764] env[62066]: DEBUG nova.compute.manager [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] No waiting events found dispatching network-vif-plugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.971923] env[62066]: WARNING nova.compute.manager [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received unexpected event network-vif-plugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 for instance with vm_state building and task_state spawning. [ 939.971963] env[62066]: DEBUG nova.compute.manager [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received event network-changed-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.972120] env[62066]: DEBUG nova.compute.manager [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Refreshing instance network info cache due to event network-changed-14b4402d-b3d0-42ef-8c2c-c02c225d8b24. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.972298] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Acquiring lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.086746] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156500, 'name': ReconfigVM_Task, 'duration_secs': 0.265793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.086746] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 940.088074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f04b7d-65d5-4e96-bcd1-cc41967986e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.111690] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.111992] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeca451e-e782-47d7-80bc-1da53be5d5cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.129828] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 940.129828] env[62066]: value = "task-1156504" [ 940.129828] env[62066]: _type = "Task" [ 940.129828] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.138091] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.204848] env[62066]: DEBUG oslo_vmware.api [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156501, 'name': SuspendVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.361896] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6a16991-555d-4df7-9052-daf03f47ea7f tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "b2afea5a-2390-40b0-b860-c50f71ebac94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.734s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.372062] env[62066]: DEBUG oslo_vmware.api [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Task: {'id': task-1156503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.428761} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.372355] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.372560] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.372802] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.373065] env[62066]: INFO nova.compute.manager [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 940.373380] env[62066]: DEBUG oslo.service.loopingcall [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.373590] env[62066]: DEBUG nova.compute.manager [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.373689] env[62066]: DEBUG nova.network.neutron [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.377653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.377947] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance network_info: |[{"id": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "address": "fa:16:3e:6a:8b:8e", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14b4402d-b3", "ovs_interfaceid": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.378487] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Acquired lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.378673] env[62066]: DEBUG nova.network.neutron [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Refreshing network info cache for port 14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 940.379878] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:8b:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14b4402d-b3d0-42ef-8c2c-c02c225d8b24', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.388417] env[62066]: DEBUG oslo.service.loopingcall [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.392023] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.392023] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0a09813-08a1-4edf-91cf-fd2e79b3f5b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.416962] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.416962] env[62066]: value = "task-1156505" [ 940.416962] env[62066]: _type = "Task" [ 940.416962] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.428426] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156505, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.641214] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156504, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.652037] env[62066]: DEBUG nova.network.neutron [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updated VIF entry in instance network info cache for port 14b4402d-b3d0-42ef-8c2c-c02c225d8b24. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.652443] env[62066]: DEBUG nova.network.neutron [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updating instance_info_cache with network_info: [{"id": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "address": "fa:16:3e:6a:8b:8e", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14b4402d-b3", "ovs_interfaceid": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.704947] env[62066]: DEBUG oslo_vmware.api [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156501, 'name': SuspendVM_Task, 'duration_secs': 0.725319} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.706654] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 940.706849] env[62066]: DEBUG nova.compute.manager [None req-86e837eb-c550-4f2e-aff2-14fbd79b5706 tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.707819] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b430b3cd-8613-49a8-9697-de7d46edfa0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.729262] env[62066]: DEBUG nova.compute.manager [req-45a7bab4-6005-4e1e-b4e8-63616ffca046 req-ab625535-e06d-45b6-a55b-f4abca87ebbd service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Received event network-vif-deleted-3e4fe3c0-de42-48eb-97b5-6f00fafd3a95 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.729262] env[62066]: INFO nova.compute.manager [req-45a7bab4-6005-4e1e-b4e8-63616ffca046 req-ab625535-e06d-45b6-a55b-f4abca87ebbd service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Neutron deleted interface 3e4fe3c0-de42-48eb-97b5-6f00fafd3a95; detaching it from the instance and deleting it from the info cache [ 940.729262] env[62066]: DEBUG nova.network.neutron [req-45a7bab4-6005-4e1e-b4e8-63616ffca046 req-ab625535-e06d-45b6-a55b-f4abca87ebbd service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.927593] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156505, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.004285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddffe29-e3ee-445f-9f7f-606ff9e6d955 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.011941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843d9ee1-ea07-46b5-97ff-50d21bd9642c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.042500] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd089cdd-7fb4-4896-83a3-6cedcc07eff4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.049918] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fd8f46-adf2-435c-b670-c80ff6a63462 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.064408] env[62066]: DEBUG nova.compute.provider_tree [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.141143] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156504, 'name': ReconfigVM_Task, 'duration_secs': 0.818149} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.141450] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f/2fc3eed6-7af2-4c34-a0d3-0498ae209f5f.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.141724] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.154862] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e1bf2b0-81eb-4af3-907b-3d6ba1a26823 req-d00ba588-8bab-4977-8c1d-332e89cbd3e6 service nova] Releasing lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.207978] env[62066]: DEBUG nova.network.neutron [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.232238] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d8606d7-5635-4f71-9c5a-cc9b904803e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.242013] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad77c8c-f0c2-43fa-94ce-6f34f374c14e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.271986] env[62066]: DEBUG nova.compute.manager [req-45a7bab4-6005-4e1e-b4e8-63616ffca046 req-ab625535-e06d-45b6-a55b-f4abca87ebbd service nova] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Detach interface failed, port_id=3e4fe3c0-de42-48eb-97b5-6f00fafd3a95, reason: Instance d50f0505-66f4-412f-9744-25e0ea96277a could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 941.427501] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156505, 'name': CreateVM_Task, 'duration_secs': 0.787291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.427686] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.428415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.428593] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.428910] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.429871] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df8fdde8-7718-4ca3-98d6-2f454eda7578 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.434220] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 941.434220] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8de34-e22e-7091-abc7-9e0407f30a37" [ 941.434220] env[62066]: _type = "Task" [ 941.434220] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.442161] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8de34-e22e-7091-abc7-9e0407f30a37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.585240] env[62066]: ERROR nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [req-34655c48-0b97-4ddd-9448-8242e920eefc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-34655c48-0b97-4ddd-9448-8242e920eefc"}]} [ 941.600235] env[62066]: DEBUG nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 941.612044] env[62066]: DEBUG nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 941.612299] env[62066]: DEBUG nova.compute.provider_tree [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.622552] env[62066]: DEBUG nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 941.637971] env[62066]: DEBUG nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 941.648045] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dcdbfc-296e-4963-b2c9-4e41f2a7f01c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.669552] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d64715d-9d9a-45bc-920f-66c0196be1ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.687644] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.710367] env[62066]: INFO nova.compute.manager [-] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Took 1.34 seconds to deallocate network for instance. [ 941.849795] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a766d16-b63d-4e60-9798-a617e281ac41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.857238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ac65e9-cbda-4110-9a92-371ea70806f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.889190] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3d5e9f-b3d2-4e83-a427-ec4c1aba9c3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.897296] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65de09c5-655d-421d-b6c6-8d9cce22a2bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.911088] env[62066]: DEBUG nova.compute.provider_tree [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 941.943933] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c8de34-e22e-7091-abc7-9e0407f30a37, 'name': SearchDatastore_Task, 'duration_secs': 0.010898} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.944365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.944618] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.944864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.945063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.945207] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.945573] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-572244ba-874b-42b5-bafc-d5a07f445cf3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.953675] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.953861] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.954814] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09f304a-8751-431c-90cc-8a7198a0c706 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.959700] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 941.959700] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521dd3f3-a722-3c7f-b79c-ec8510b6aa86" [ 941.959700] env[62066]: _type = "Task" [ 941.959700] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.968010] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521dd3f3-a722-3c7f-b79c-ec8510b6aa86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.974806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.975044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.216721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.235600] env[62066]: DEBUG nova.network.neutron [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Port aef860e5-c337-4d8f-b0ff-e3c92e1ff75f binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 942.308391] env[62066]: INFO nova.compute.manager [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Resuming [ 942.309020] env[62066]: DEBUG nova.objects.instance [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lazy-loading 'flavor' on Instance uuid 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.445888] env[62066]: DEBUG nova.scheduler.client.report [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 942.446252] env[62066]: DEBUG nova.compute.provider_tree [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 118 to 119 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 942.447350] env[62066]: DEBUG nova.compute.provider_tree [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 942.471127] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521dd3f3-a722-3c7f-b79c-ec8510b6aa86, 'name': SearchDatastore_Task, 'duration_secs': 0.008442} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.471964] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-472980ea-d3ad-4b45-b24b-684d33a6c840 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.476977] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.480178] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 942.480178] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bb4813-96a8-1666-f326-d3d3ca3e4efd" [ 942.480178] env[62066]: _type = "Task" [ 942.480178] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.489113] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bb4813-96a8-1666-f326-d3d3ca3e4efd, 'name': SearchDatastore_Task, 'duration_secs': 0.009182} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.489374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.489637] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3e7a2dd5-f01e-4edd-884f-7d614f260181/3e7a2dd5-f01e-4edd-884f-7d614f260181.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.489894] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbfe7e36-8ac4-49ce-a38a-6200b7dd5329 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.496764] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 942.496764] env[62066]: value = "task-1156506" [ 942.496764] env[62066]: _type = "Task" [ 942.496764] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.506417] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.952836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.125s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.952836] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.955735] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.270s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.955735] env[62066]: DEBUG nova.objects.instance [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'resources' on Instance uuid f24f8067-07b2-4941-8464-c30b58983ec1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.001410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.009144] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440998} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.009460] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3e7a2dd5-f01e-4edd-884f-7d614f260181/3e7a2dd5-f01e-4edd-884f-7d614f260181.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.009755] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.010423] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1319e7bc-10f2-4ce0-86d2-c6d9a38fa1bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.016744] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 943.016744] env[62066]: value = "task-1156507" [ 943.016744] env[62066]: _type = "Task" [ 943.016744] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.024556] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.259811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.260204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.260251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.317015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.317226] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquired lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.317506] env[62066]: DEBUG nova.network.neutron [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.459636] env[62066]: DEBUG nova.compute.utils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 943.464827] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.465030] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 943.527926] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061656} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.530993] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.532249] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8a03a6-781b-4008-8276-63e6dfa0274b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.556031] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 3e7a2dd5-f01e-4edd-884f-7d614f260181/3e7a2dd5-f01e-4edd-884f-7d614f260181.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.557546] env[62066]: DEBUG nova.policy [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91c19d14b9af4bc68ba1f604b117efd9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4763071af33b435ba0b7f94b7dcb4de2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.561395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f5aef70-cc98-4bbb-8d7e-3919f295aab9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.585285] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 943.585285] env[62066]: value = "task-1156508" [ 943.585285] env[62066]: _type = "Task" [ 943.585285] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.593809] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.686013] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e904d7-d6b3-4bd9-9cd5-2c8d1c7721d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.694070] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a206b2cb-51f5-4eec-b64e-0cbcf570407e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.724906] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd4d03b-4ce6-4c25-ba89-c4457f04c233 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.733478] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d392180-84e1-4195-9409-20fec8087b2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.747901] env[62066]: DEBUG nova.compute.provider_tree [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.927298] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Successfully created port: f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.965803] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 944.042698] env[62066]: DEBUG nova.network.neutron [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [{"id": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "address": "fa:16:3e:7c:3e:99", "network": {"id": "7110df11-8df7-4b7a-80d4-69d312ba01e9", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1444034210-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61d04b8fe994dcfaa361bb87eb587d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e02e98f-44ce-42b7-a3ac-4034fae5d127", "external-id": "nsx-vlan-transportzone-874", "segmentation_id": 874, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9795c0a-c0", "ovs_interfaceid": "c9795c0a-c035-447a-9433-6cd7ff5ab2ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.095735] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156508, 'name': ReconfigVM_Task, 'duration_secs': 0.295282} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.096040] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 3e7a2dd5-f01e-4edd-884f-7d614f260181/3e7a2dd5-f01e-4edd-884f-7d614f260181.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.096678] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b127754b-45f0-4d52-b0c1-d25ccf34bad7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.103645] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 944.103645] env[62066]: value = "task-1156509" [ 944.103645] env[62066]: _type = "Task" [ 944.103645] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.113133] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156509, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.250670] env[62066]: DEBUG nova.scheduler.client.report [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.298693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.298946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.299118] env[62066]: DEBUG nova.network.neutron [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.546031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Releasing lock "refresh_cache-0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.547222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f683b8a5-dd74-45ab-9bbe-770013bf1a9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.554822] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Resuming the VM {{(pid=62066) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 944.555120] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-beaf80e1-51b2-43a7-b011-75550689c4d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.563641] env[62066]: DEBUG oslo_vmware.api [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 944.563641] env[62066]: value = "task-1156510" [ 944.563641] env[62066]: _type = "Task" [ 944.563641] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.571317] env[62066]: DEBUG oslo_vmware.api [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.613725] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156509, 'name': Rename_Task, 'duration_secs': 0.141901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.614037] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.614307] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eccb266a-1a3b-4ede-b688-25a2fc050e34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.621538] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 944.621538] env[62066]: value = "task-1156511" [ 944.621538] env[62066]: _type = "Task" [ 944.621538] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.629793] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156511, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.756514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.758987] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.542s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.759261] env[62066]: DEBUG nova.objects.instance [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lazy-loading 'resources' on Instance uuid d50f0505-66f4-412f-9744-25e0ea96277a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.781076] env[62066]: INFO nova.scheduler.client.report [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted allocations for instance f24f8067-07b2-4941-8464-c30b58983ec1 [ 944.978290] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.007657] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.007941] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.008126] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.008317] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.008470] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.008622] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.008835] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.008998] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.009284] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.009413] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.009613] env[62066]: DEBUG nova.virt.hardware [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.010588] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850e66b6-3702-42a6-9463-fb02f100d76a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.014452] env[62066]: DEBUG nova.network.neutron [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.026859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8cae80-edd4-4419-adba-443f28311516 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.078622] env[62066]: DEBUG oslo_vmware.api [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156510, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.135441] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156511, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.294019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-157fb202-4f7b-4b96-a21e-c288f6753845 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "f24f8067-07b2-4941-8464-c30b58983ec1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.907s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.447144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8750c872-c573-464f-87f3-aa5a00ab5909 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.458172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cb2f9c-7476-46b0-ab92-933ae7e33e8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.493690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385213b4-5945-4742-b9b8-b208eeec7387 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.502517] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7476f053-b58a-4515-82a4-dfe03dae9850 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.520127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.524578] env[62066]: DEBUG nova.compute.provider_tree [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.576845] env[62066]: DEBUG oslo_vmware.api [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156510, 'name': PowerOnVM_Task, 'duration_secs': 0.542266} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.577332] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Resumed the VM {{(pid=62066) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 945.577617] env[62066]: DEBUG nova.compute.manager [None req-feffce0d-50e7-47b5-8007-0368ce88183f tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.578906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.579220] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.581019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2595ed-1e76-4569-813f-6f89c84bcc69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.633842] env[62066]: DEBUG oslo_vmware.api [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156511, 'name': PowerOnVM_Task, 'duration_secs': 0.562528} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.634274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.634555] env[62066]: INFO nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Took 7.59 seconds to spawn the instance on the hypervisor. [ 945.634803] env[62066]: DEBUG nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.635637] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e115f9a6-d731-45de-a170-3c1f2dc71d8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.698764] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.876505] env[62066]: DEBUG nova.compute.manager [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Received event network-vif-plugged-f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.876505] env[62066]: DEBUG oslo_concurrency.lockutils [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] Acquiring lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.876505] env[62066]: DEBUG oslo_concurrency.lockutils [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] Lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.878995] env[62066]: DEBUG oslo_concurrency.lockutils [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] Lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.878995] env[62066]: DEBUG nova.compute.manager [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] No waiting events found dispatching network-vif-plugged-f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.878995] env[62066]: WARNING nova.compute.manager [req-fef0ecb2-4c35-4a83-9503-256f6c2dc5f3 req-2cabc070-bad0-4217-b0b0-49bc737d21e4 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Received unexpected event network-vif-plugged-f414725f-8af3-4a19-acf5-6cc44ff37ecb for instance with vm_state building and task_state spawning. [ 946.030251] env[62066]: DEBUG nova.scheduler.client.report [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.051238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a15300-5e24-4cab-9d27-212e304cb898 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.073705] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47679ba0-c47e-41f6-ad2b-5c689ef47736 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.082454] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 946.087533] env[62066]: DEBUG nova.compute.utils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 946.153511] env[62066]: INFO nova.compute.manager [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Took 15.20 seconds to build instance. [ 946.201847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.371471] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Successfully updated port: f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 946.402881] env[62066]: DEBUG nova.compute.manager [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Received event network-changed-f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.403273] env[62066]: DEBUG nova.compute.manager [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Refreshing instance network info cache due to event network-changed-f414725f-8af3-4a19-acf5-6cc44ff37ecb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.403549] env[62066]: DEBUG oslo_concurrency.lockutils [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] Acquiring lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.403756] env[62066]: DEBUG oslo_concurrency.lockutils [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] Acquired lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.403980] env[62066]: DEBUG nova.network.neutron [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Refreshing network info cache for port f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.536893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.543697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.541s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.544438] env[62066]: INFO nova.compute.claims [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.569615] env[62066]: INFO nova.scheduler.client.report [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Deleted allocations for instance d50f0505-66f4-412f-9744-25e0ea96277a [ 946.593181] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.595967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.595967] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c5ff8f6-8e79-4202-b90e-6eaf0e9b7fe8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.606028] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 946.606028] env[62066]: value = "task-1156512" [ 946.606028] env[62066]: _type = "Task" [ 946.606028] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.615626] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.655620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a968c316-1c68-4fec-b87b-b0d3b2786bce tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.716s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.874403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.947632] env[62066]: DEBUG nova.network.neutron [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.048671] env[62066]: DEBUG nova.network.neutron [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.081353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6ba2ab0-9399-4907-b703-2c9c6e0c5e26 tempest-ServerRescueNegativeTestJSON-904489895 tempest-ServerRescueNegativeTestJSON-904489895-project-member] Lock "d50f0505-66f4-412f-9744-25e0ea96277a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.829s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.116695] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156512, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.552951] env[62066]: DEBUG oslo_concurrency.lockutils [req-004ed4e9-b106-447c-b472-21b85cb6f56b req-2f7deed1-33fb-42db-8b7e-1ee15edf9ea1 service nova] Releasing lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.556281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquired lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.556539] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.617568] env[62066]: DEBUG oslo_vmware.api [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156512, 'name': PowerOnVM_Task, 'duration_secs': 0.71655} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.617856] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.618061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-13551adb-3596-4fd4-91e5-e0b4ad02c290 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance '2fc3eed6-7af2-4c34-a0d3-0498ae209f5f' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.656015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.656015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.656015] env[62066]: INFO nova.compute.manager [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Attaching volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 to /dev/sdb [ 947.701026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90eab61-2bff-4ed1-a38c-06e316724c94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.711554] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c9daa1-956f-4906-a0b8-e356b1f6e0f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.718471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.718734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.733275] env[62066]: DEBUG nova.virt.block_device [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating existing volume attachment record: d6d344b3-f7ed-46dc-822a-24fb67c039d6 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 947.767502] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f32b49-add4-45fd-bc0c-116603bfc794 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.776185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca920ce-546b-427b-9869-9fd31e44d6ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.811207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006cd740-38ca-456b-aeb3-b896c0fc40b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.820575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c115ae-5d41-4c85-9a7b-c1f855d7bdfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.837571] env[62066]: DEBUG nova.compute.provider_tree [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.131185] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.131723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.135018] env[62066]: INFO nova.compute.manager [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Shelving [ 948.135018] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.221665] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.342978] env[62066]: DEBUG nova.scheduler.client.report [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.515241] env[62066]: DEBUG nova.network.neutron [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updating instance_info_cache with network_info: [{"id": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "address": "fa:16:3e:62:db:6b", "network": {"id": "8090756b-78c4-4068-a3d0-489495be55e0", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-321897902-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4763071af33b435ba0b7f94b7dcb4de2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf414725f-8a", "ovs_interfaceid": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.645433] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.645433] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d4e2075-248a-4cd2-a4f2-29355e37e5e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.655305] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 948.655305] env[62066]: value = "task-1156516" [ 948.655305] env[62066]: _type = "Task" [ 948.655305] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.666036] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.758993] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.847035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.847654] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.854025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.649s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.854025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.854025] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 948.854025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.094s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.854025] env[62066]: INFO nova.compute.claims [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.856759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc8c3e8-03e5-44b4-87e5-e24dec80c296 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.870047] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1784dee6-6572-427c-96bf-0f469d5ff5c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.887759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9b54d2-efa4-42f0-b0b5-9a95326b7f5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.897997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89809731-e576-4fd5-acb2-69c5452528ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.931425] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179760MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 948.931711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.017607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Releasing lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.018138] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance network_info: |[{"id": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "address": "fa:16:3e:62:db:6b", "network": {"id": "8090756b-78c4-4068-a3d0-489495be55e0", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-321897902-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4763071af33b435ba0b7f94b7dcb4de2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf414725f-8a", "ovs_interfaceid": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.018752] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:db:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4055505f-97ab-400b-969c-43d99b38fd48', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f414725f-8af3-4a19-acf5-6cc44ff37ecb', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.027379] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Creating folder: Project (4763071af33b435ba0b7f94b7dcb4de2). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.027688] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e9ccf64-f831-45f0-b2f8-936de40e7dba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.042161] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Created folder: Project (4763071af33b435ba0b7f94b7dcb4de2) in parent group-v251573. [ 949.042254] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Creating folder: Instances. Parent ref: group-v251749. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.042596] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56fcc125-c054-403f-9763-1c9a931b9a9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.054937] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Created folder: Instances in parent group-v251749. [ 949.055224] env[62066]: DEBUG oslo.service.loopingcall [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.055477] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.055690] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c71e6e38-dbaa-46fe-80a1-51555ec71e20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.078451] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.078451] env[62066]: value = "task-1156519" [ 949.078451] env[62066]: _type = "Task" [ 949.078451] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.087956] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156519, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.168632] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156516, 'name': PowerOffVM_Task, 'duration_secs': 0.226163} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.169095] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.170487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ddca35-fd8b-46b3-9c48-7ef77f629514 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.198504] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c506416-f720-4e1b-8f4f-70153a67da09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.357827] env[62066]: DEBUG nova.compute.utils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 949.359374] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.359549] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.490185] env[62066]: DEBUG nova.policy [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.594361] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156519, 'name': CreateVM_Task, 'duration_secs': 0.457124} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.594604] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.595327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.595504] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.595831] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.597559] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de36d59a-37a6-4d64-b9b9-fb345a2d71bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.606770] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 949.606770] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]526ad2dd-898d-8c87-cf0a-1da320ac1285" [ 949.606770] env[62066]: _type = "Task" [ 949.606770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.616385] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526ad2dd-898d-8c87-cf0a-1da320ac1285, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.711248] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 949.711625] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7c8b3223-71d0-4e89-8d9c-3c8f86aefc28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.721729] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 949.721729] env[62066]: value = "task-1156520" [ 949.721729] env[62066]: _type = "Task" [ 949.721729] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.731695] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156520, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.863528] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.078797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec88edd5-6706-434c-bfae-929633d56402 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.087783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1559acca-a7bf-40b3-9581-2427ec2bc4f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.126529] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07d042a-d166-4334-aa03-e781b5a4d356 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.143620] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]526ad2dd-898d-8c87-cf0a-1da320ac1285, 'name': SearchDatastore_Task, 'duration_secs': 0.011207} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.147432] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148ec38d-d28a-407e-8337-ea187277b90b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.152256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.152685] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.153102] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.153437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.153774] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.154984] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b43eb1d4-1324-43a2-a1b1-704315000bc7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.169697] env[62066]: DEBUG nova.compute.provider_tree [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.171597] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.171771] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.172677] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86e94f0b-d9cd-4727-befd-3392891c9741 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.179294] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 950.179294] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52918d0a-d255-bb57-73d4-f6116d7460f2" [ 950.179294] env[62066]: _type = "Task" [ 950.179294] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.189723] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52918d0a-d255-bb57-73d4-f6116d7460f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.232426] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156520, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.531926] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Successfully created port: ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.644991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.645389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.645573] env[62066]: DEBUG nova.compute.manager [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Going to confirm migration 4 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 950.674372] env[62066]: DEBUG nova.scheduler.client.report [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.700811] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52918d0a-d255-bb57-73d4-f6116d7460f2, 'name': SearchDatastore_Task, 'duration_secs': 0.013002} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.701727] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2092c643-0242-4465-8ae0-cd571999969b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.710552] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 950.710552] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e4913d-2ab9-8ca0-17de-c03e01a4aae2" [ 950.710552] env[62066]: _type = "Task" [ 950.710552] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.722121] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e4913d-2ab9-8ca0-17de-c03e01a4aae2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.731530] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156520, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.877366] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.914996] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.915279] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.915521] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.915628] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.915778] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.915928] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.916152] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.916325] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.916499] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.916664] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.916837] env[62066]: DEBUG nova.virt.hardware [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.921707] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e4a73c-11ce-47cc-ac4e-578530081bc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.934628] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce339edb-9ed7-41c5-863e-d41019dcf30f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.182034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.182955] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.190874] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.259s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.228086] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e4913d-2ab9-8ca0-17de-c03e01a4aae2, 'name': SearchDatastore_Task, 'duration_secs': 0.015007} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.234326] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.234988] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 16390d89-f2d1-4d6c-be62-b85e61906865/16390d89-f2d1-4d6c-be62-b85e61906865.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.235312] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-281995a9-4490-4003-9b32-fc9eaa0941d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.245356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.245590] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.246388] env[62066]: DEBUG nova.network.neutron [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 951.246511] env[62066]: DEBUG nova.objects.instance [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'info_cache' on Instance uuid 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.258518] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156520, 'name': CreateSnapshot_Task, 'duration_secs': 1.092643} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.258883] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 951.258883] env[62066]: value = "task-1156522" [ 951.258883] env[62066]: _type = "Task" [ 951.258883] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.259217] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 951.260728] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f903379-76d5-49ed-a274-9140b9356141 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.280351] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.692543] env[62066]: DEBUG nova.compute.utils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 951.694204] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 951.694435] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 951.750717] env[62066]: DEBUG nova.policy [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.771678] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156522, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.790147] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 951.790147] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d5616fa9-88fb-4509-8ae9-c7e4c5b88e47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.801386] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 951.801386] env[62066]: value = "task-1156523" [ 951.801386] env[62066]: _type = "Task" [ 951.801386] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.812691] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156523, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.099547] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Successfully created port: 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.186665] env[62066]: DEBUG nova.compute.manager [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Received event network-vif-plugged-ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.186941] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] Acquiring lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.187127] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.187305] env[62066]: DEBUG oslo_concurrency.lockutils [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.187481] env[62066]: DEBUG nova.compute.manager [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] No waiting events found dispatching network-vif-plugged-ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 952.187653] env[62066]: WARNING nova.compute.manager [req-d8145eb3-2e94-4b9c-8b7a-7e09759e57f5 req-b25e2e3f-da7d-4043-acf7-efa5325f6cbe service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Received unexpected event network-vif-plugged-ef31c8ff-7052-4009-9b1c-64674449bae4 for instance with vm_state building and task_state spawning. [ 952.197369] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.204347] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Applying migration context for instance 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f as it has an incoming, in-progress migration 7806705f-7018-4c75-92eb-f4d8a764f5ed. Migration status is confirming {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 952.205853] env[62066]: INFO nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating resource usage from migration 7806705f-7018-4c75-92eb-f4d8a764f5ed [ 952.230831] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 212dac6a-a291-4ca8-87fb-97ebcca7976c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231057] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231230] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 31e50f97-f873-44ad-9923-67923cdb8d3a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231489] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 00f44ecb-768a-4db3-b229-27bb6f27ede1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231569] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f82c8bf2-24f3-4732-a56d-1bddaef14158 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231652] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231871] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.231971] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Migration 7806705f-7018-4c75-92eb-f4d8a764f5ed is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 952.232046] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.232194] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3e7a2dd5-f01e-4edd-884f-7d614f260181 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.232306] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 16390d89-f2d1-4d6c-be62-b85e61906865 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.232508] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 8bc2c334-ff46-44b2-b31a-d85273ec9a45 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.232579] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6e0065b0-da4f-4288-a858-8648572f8148 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 952.232788] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 952.232952] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 952.272420] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.758055} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.275276] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 16390d89-f2d1-4d6c-be62-b85e61906865/16390d89-f2d1-4d6c-be62-b85e61906865.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.275541] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.276189] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19fa08e5-1afe-4fa0-bf5f-c97c35c90cfe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.284939] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 952.284939] env[62066]: value = "task-1156524" [ 952.284939] env[62066]: _type = "Task" [ 952.284939] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.302549] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.314454] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156523, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.351123] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Successfully updated port: ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.530501] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b597a8-5151-4f28-9a0d-e5a3e5cd176c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.538594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1cab15-1c6e-491f-86de-709b4527c96f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.543041] env[62066]: DEBUG nova.network.neutron [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [{"id": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "address": "fa:16:3e:5d:ef:ab", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaef860e5-c3", "ovs_interfaceid": "aef860e5-c337-4d8f-b0ff-e3c92e1ff75f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.571487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c4b60b-0c6e-493a-9dba-4e393bfe9df1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.580221] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ab1240-40c5-46b9-a0b3-667bd15cf0ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.595639] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.782458] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 952.782753] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 952.783779] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab08e4a-4232-4932-ab5f-19a240305b7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.797272] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07907} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.807786] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.809091] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4562aa-252e-4ec9-89a6-bfc466143ad2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.814668] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d3d186-9001-4ffe-b60e-0a3a5022c9d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.822710] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156523, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.851633] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.859983] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 16390d89-f2d1-4d6c-be62-b85e61906865/16390d89-f2d1-4d6c-be62-b85e61906865.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.860295] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6544241b-c584-4a68-8776-06c0ca0e7c23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.873465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.873465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.873465] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.874577] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d05355c0-b28a-4581-add3-4737c615f2f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.896205] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 952.896205] env[62066]: value = "task-1156525" [ 952.896205] env[62066]: _type = "Task" [ 952.896205] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.898578] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 952.898578] env[62066]: value = "task-1156526" [ 952.898578] env[62066]: _type = "Task" [ 952.898578] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.912711] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156526, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.912955] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156525, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.045549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.045805] env[62066]: DEBUG nova.objects.instance [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'migration_context' on Instance uuid 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.117629] env[62066]: ERROR nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [req-6afcadfd-65f4-4fd6-80c1-70d2c9fd8779] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6afcadfd-65f4-4fd6-80c1-70d2c9fd8779"}]} [ 953.136586] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 953.207580] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.236289] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.236637] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.236811] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.237030] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.237218] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.237401] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.237664] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.237847] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.238066] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.238264] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.238465] env[62066]: DEBUG nova.virt.hardware [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.239407] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d73acb7-f78a-4874-a998-ffe4391ed713 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.242960] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 953.243367] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.253479] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad844333-57c2-408d-9efa-9eb9c0d0fe65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.260628] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 953.279439] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 953.322400] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156523, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.412333] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156525, 'name': ReconfigVM_Task, 'duration_secs': 0.329454} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.418642] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 16390d89-f2d1-4d6c-be62-b85e61906865/16390d89-f2d1-4d6c-be62-b85e61906865.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.419379] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156526, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.420039] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6f58dd2-c010-4679-a94d-f9c45eae6588 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.422327] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 953.432296] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 953.432296] env[62066]: value = "task-1156527" [ 953.432296] env[62066]: _type = "Task" [ 953.432296] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.441464] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156527, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.472522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be02a59f-ace4-49a0-b4fc-ea77ac09e246 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.486333] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0fb4048-e75f-4f7c-865e-a26167c5893e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.518257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0b1fcd-2b07-476d-b7d5-5f984bbb76c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.526682] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1424d14e-412b-42a4-a837-2073f9d7de65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.541968] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.550157] env[62066]: DEBUG nova.objects.base [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Object Instance<2fc3eed6-7af2-4c34-a0d3-0498ae209f5f> lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 953.552122] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd613867-6570-4ee8-99d2-98b6a4d6d1f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.574316] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5557c3fc-810e-413c-9165-59eb41f941df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.580467] env[62066]: DEBUG oslo_vmware.api [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 953.580467] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523b4881-4a6d-327e-04d4-ccd2fa3f104f" [ 953.580467] env[62066]: _type = "Task" [ 953.580467] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.589698] env[62066]: DEBUG oslo_vmware.api [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523b4881-4a6d-327e-04d4-ccd2fa3f104f, 'name': SearchDatastore_Task, 'duration_secs': 0.007011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.589981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.665604] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.665886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.666140] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.666341] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.666516] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.668793] env[62066]: INFO nova.compute.manager [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Terminating instance [ 953.670666] env[62066]: DEBUG nova.compute.manager [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.670860] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.671784] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80a6129-366a-47ab-9e3d-214c681fecca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.675264] env[62066]: DEBUG nova.network.neutron [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Updating instance_info_cache with network_info: [{"id": "ef31c8ff-7052-4009-9b1c-64674449bae4", "address": "fa:16:3e:24:ce:ed", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef31c8ff-70", "ovs_interfaceid": "ef31c8ff-7052-4009-9b1c-64674449bae4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.681505] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.683492] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-648bf9f8-be68-455f-8882-6986e0ec3f37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.690245] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 953.690245] env[62066]: value = "task-1156528" [ 953.690245] env[62066]: _type = "Task" [ 953.690245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.700023] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.711986] env[62066]: DEBUG nova.compute.manager [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-vif-plugged-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.712293] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.712602] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.713025] env[62066]: DEBUG oslo_concurrency.lockutils [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.713155] env[62066]: DEBUG nova.compute.manager [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] No waiting events found dispatching network-vif-plugged-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.713358] env[62066]: WARNING nova.compute.manager [req-ef458912-7cb7-487e-be27-8d06f3644530 req-9013776c-81ee-4581-90f0-ffa3fdec9997 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received unexpected event network-vif-plugged-00ffbc69-586f-4442-b0fc-616ec3c6c49b for instance with vm_state building and task_state spawning. [ 953.801231] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Successfully updated port: 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.824091] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156523, 'name': CloneVM_Task, 'duration_secs': 1.931886} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.824386] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Created linked-clone VM from snapshot [ 953.825175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1e9a47-c824-442c-bbe8-ef93f9563fd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.833627] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Uploading image 1b87e52b-3c5d-4619-b6c5-f6075eb5ef9b {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 953.860573] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 953.860573] env[62066]: value = "vm-251753" [ 953.860573] env[62066]: _type = "VirtualMachine" [ 953.860573] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 953.860869] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-446afe0a-6e4f-456e-a781-e59cae53a9be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.868135] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lease: (returnval){ [ 953.868135] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527b117a-70af-1288-d26e-ecfacc064924" [ 953.868135] env[62066]: _type = "HttpNfcLease" [ 953.868135] env[62066]: } obtained for exporting VM: (result){ [ 953.868135] env[62066]: value = "vm-251753" [ 953.868135] env[62066]: _type = "VirtualMachine" [ 953.868135] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 953.868438] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the lease: (returnval){ [ 953.868438] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527b117a-70af-1288-d26e-ecfacc064924" [ 953.868438] env[62066]: _type = "HttpNfcLease" [ 953.868438] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.875452] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.875452] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527b117a-70af-1288-d26e-ecfacc064924" [ 953.875452] env[62066]: _type = "HttpNfcLease" [ 953.875452] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.912211] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156526, 'name': ReconfigVM_Task, 'duration_secs': 0.678286} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.912499] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.917472] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5b5ab35-c16d-4c2f-9d71-768c37d4190e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.936062] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 953.936062] env[62066]: value = "task-1156530" [ 953.936062] env[62066]: _type = "Task" [ 953.936062] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.943091] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156527, 'name': Rename_Task, 'duration_secs': 0.160792} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.943797] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.944060] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbdf4428-35ee-4308-ad85-d8949808a474 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.948853] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156530, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.954446] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 953.954446] env[62066]: value = "task-1156531" [ 953.954446] env[62066]: _type = "Task" [ 953.954446] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.963594] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.077083] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 954.077347] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 120 to 121 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 954.077523] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.178096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.178429] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Instance network_info: |[{"id": "ef31c8ff-7052-4009-9b1c-64674449bae4", "address": "fa:16:3e:24:ce:ed", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef31c8ff-70", "ovs_interfaceid": "ef31c8ff-7052-4009-9b1c-64674449bae4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.178877] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:ce:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef31c8ff-7052-4009-9b1c-64674449bae4', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.186863] env[62066]: DEBUG oslo.service.loopingcall [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.187114] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.187795] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73020d47-fbab-4714-9564-69f8e666cb01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.212545] env[62066]: DEBUG nova.compute.manager [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Received event network-changed-ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.212780] env[62066]: DEBUG nova.compute.manager [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Refreshing instance network info cache due to event network-changed-ef31c8ff-7052-4009-9b1c-64674449bae4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 954.213064] env[62066]: DEBUG oslo_concurrency.lockutils [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] Acquiring lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.213321] env[62066]: DEBUG oslo_concurrency.lockutils [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] Acquired lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.213559] env[62066]: DEBUG nova.network.neutron [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Refreshing network info cache for port ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.219629] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156528, 'name': PowerOffVM_Task, 'duration_secs': 0.175785} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.219764] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.219764] env[62066]: value = "task-1156532" [ 954.219764] env[62066]: _type = "Task" [ 954.219764] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.220234] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.220423] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.221028] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cc1fa6c-b4b8-426c-b611-b0a96c234f9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.233181] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156532, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.306450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.306450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.306450] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.310164] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.310396] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.310658] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleting the datastore file [datastore1] 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.310857] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-146e7a71-b9de-4539-acd0-435df2a320ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.321041] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for the task: (returnval){ [ 954.321041] env[62066]: value = "task-1156534" [ 954.321041] env[62066]: _type = "Task" [ 954.321041] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.329406] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.378110] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.378110] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527b117a-70af-1288-d26e-ecfacc064924" [ 954.378110] env[62066]: _type = "HttpNfcLease" [ 954.378110] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.378475] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.378475] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527b117a-70af-1288-d26e-ecfacc064924" [ 954.378475] env[62066]: _type = "HttpNfcLease" [ 954.378475] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.379293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e537a35-26be-467b-94c5-c3bbe93d5fb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.388166] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.388407] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 954.461795] env[62066]: DEBUG oslo_vmware.api [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156530, 'name': ReconfigVM_Task, 'duration_secs': 0.156099} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.462625] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 954.467990] env[62066]: DEBUG oslo_vmware.api [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156531, 'name': PowerOnVM_Task, 'duration_secs': 0.467691} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.468561] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.468830] env[62066]: INFO nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Took 9.49 seconds to spawn the instance on the hypervisor. [ 954.469065] env[62066]: DEBUG nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.469967] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6af52dd-4686-45c3-8fea-f4ca36e5c8a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.488976] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-08241503-f2f5-47f7-b2e4-eb93842d6f9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.583569] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 954.583788] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.393s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.584089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.994s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.733070] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156532, 'name': CreateVM_Task, 'duration_secs': 0.418502} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.733287] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.733986] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.734195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.734540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.734802] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae7828b5-0c59-4ae2-b650-490deec55b2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.740687] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 954.740687] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c850b6-f078-bad1-6245-6c53b938e05a" [ 954.740687] env[62066]: _type = "Task" [ 954.740687] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.751305] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c850b6-f078-bad1-6245-6c53b938e05a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.831271] env[62066]: DEBUG oslo_vmware.api [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Task: {'id': task-1156534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201271} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.831812] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.831812] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.831962] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.832150] env[62066]: INFO nova.compute.manager [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Took 1.16 seconds to destroy the instance on the hypervisor. [ 954.832838] env[62066]: DEBUG oslo.service.loopingcall [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.832916] env[62066]: DEBUG nova.compute.manager [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.833394] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 954.857508] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 954.991483] env[62066]: INFO nova.compute.manager [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Took 17.12 seconds to build instance. [ 955.076622] env[62066]: DEBUG nova.network.neutron [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Updated VIF entry in instance network info cache for port ef31c8ff-7052-4009-9b1c-64674449bae4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 955.077070] env[62066]: DEBUG nova.network.neutron [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Updating instance_info_cache with network_info: [{"id": "ef31c8ff-7052-4009-9b1c-64674449bae4", "address": "fa:16:3e:24:ce:ed", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef31c8ff-70", "ovs_interfaceid": "ef31c8ff-7052-4009-9b1c-64674449bae4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.095303] env[62066]: DEBUG nova.network.neutron [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.255808] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c850b6-f078-bad1-6245-6c53b938e05a, 'name': SearchDatastore_Task, 'duration_secs': 0.014614} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.258389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.258687] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.258939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.259749] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.259749] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.260625] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9ada1c3-7619-481f-b980-0be7a842977d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.271379] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.271638] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.275011] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ee2d1f4-3a47-45e0-b94d-1e437e13dcb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.281880] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 955.281880] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a30e46-286e-d233-480e-c29521cd5024" [ 955.281880] env[62066]: _type = "Task" [ 955.281880] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.290645] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a30e46-286e-d233-480e-c29521cd5024, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.308704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40d7d2e-677c-4f4c-8795-f3691705f451 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.319158] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586a4fec-beea-4186-93e8-169e027e60bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.354382] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353e8f55-48df-4626-bbcb-0d64c8242505 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.363237] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365795d5-196d-4c57-8441-402feea28fe0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.380762] env[62066]: DEBUG nova.compute.provider_tree [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.494312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3559675-0f72-41dd-a161-d186a5a6ec25 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.647s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.522813] env[62066]: DEBUG nova.objects.instance [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'flavor' on Instance uuid 00f44ecb-768a-4db3-b229-27bb6f27ede1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.580270] env[62066]: DEBUG oslo_concurrency.lockutils [req-48c48383-b028-4220-bb4b-db866b457da6 req-1a9a14c2-cf9c-4b6d-a6e3-eeb05f3a8c64 service nova] Releasing lock "refresh_cache-8bc2c334-ff46-44b2-b31a-d85273ec9a45" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.599169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.599509] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Instance network_info: |[{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.599964] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:aa:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '65497291-07f3-434c-bd42-657a0cb03365', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00ffbc69-586f-4442-b0fc-616ec3c6c49b', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.608661] env[62066]: DEBUG oslo.service.loopingcall [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.608661] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.608661] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82eb6b81-1d9c-4faa-9124-53677364a8f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.630861] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.630861] env[62066]: value = "task-1156535" [ 955.630861] env[62066]: _type = "Task" [ 955.630861] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.640496] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156535, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.793492] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a30e46-286e-d233-480e-c29521cd5024, 'name': SearchDatastore_Task, 'duration_secs': 0.019712} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.794400] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-871c74b1-de41-4c88-bc46-979af6f9fdf1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.796984] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.802475] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 955.802475] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8efcf-1f94-c417-51bb-b68bfd9a714a" [ 955.802475] env[62066]: _type = "Task" [ 955.802475] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.812778] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8efcf-1f94-c417-51bb-b68bfd9a714a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.843385] env[62066]: DEBUG nova.compute.manager [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.843686] env[62066]: DEBUG nova.compute.manager [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.843686] env[62066]: DEBUG oslo_concurrency.lockutils [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.843830] env[62066]: DEBUG oslo_concurrency.lockutils [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.844032] env[62066]: DEBUG nova.network.neutron [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.883507] env[62066]: DEBUG nova.scheduler.client.report [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.030060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-211c1878-0988-4cf2-ac1d-b95d60add824 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.374s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.142660] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156535, 'name': CreateVM_Task, 'duration_secs': 0.353014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.142858] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.143600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.143767] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.144097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.144368] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bffa5f3a-5c00-43d0-bd66-58e82338c3b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.152882] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 956.152882] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e17895-b7a1-127f-5c84-556e79d22a09" [ 956.152882] env[62066]: _type = "Task" [ 956.152882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.169192] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e17895-b7a1-127f-5c84-556e79d22a09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.243854] env[62066]: DEBUG nova.compute.manager [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Received event network-vif-deleted-c9795c0a-c035-447a-9433-6cd7ff5ab2ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.244253] env[62066]: DEBUG nova.compute.manager [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Received event network-changed-f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.244448] env[62066]: DEBUG nova.compute.manager [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Refreshing instance network info cache due to event network-changed-f414725f-8af3-4a19-acf5-6cc44ff37ecb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.244741] env[62066]: DEBUG oslo_concurrency.lockutils [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] Acquiring lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.244897] env[62066]: DEBUG oslo_concurrency.lockutils [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] Acquired lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.245215] env[62066]: DEBUG nova.network.neutron [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Refreshing network info cache for port f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.300069] env[62066]: INFO nova.compute.manager [-] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Took 1.47 seconds to deallocate network for instance. [ 956.315842] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8efcf-1f94-c417-51bb-b68bfd9a714a, 'name': SearchDatastore_Task, 'duration_secs': 0.012621} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.315917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.316129] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 8bc2c334-ff46-44b2-b31a-d85273ec9a45/8bc2c334-ff46-44b2-b31a-d85273ec9a45.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 956.316394] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e87bc44e-e728-406f-b1ac-cc4f2c79a7f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.324034] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 956.324034] env[62066]: value = "task-1156536" [ 956.324034] env[62066]: _type = "Task" [ 956.324034] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.334477] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.557299] env[62066]: INFO nova.compute.manager [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Rebuilding instance [ 956.621502] env[62066]: DEBUG nova.compute.manager [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.626207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad81024-6005-4baa-a66a-73322fd9883c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.666017] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e17895-b7a1-127f-5c84-556e79d22a09, 'name': SearchDatastore_Task, 'duration_secs': 0.023264} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.666017] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.666017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.666448] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.666527] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.666763] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.667752] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47651845-9e2e-4e12-bee7-0402e13ddaf8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.677441] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.677636] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.678560] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-197b8ca0-de12-49dd-be12-7d61763c698a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.685128] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 956.685128] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c16734-81c8-f269-d1b5-e32181686f6d" [ 956.685128] env[62066]: _type = "Task" [ 956.685128] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.694468] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c16734-81c8-f269-d1b5-e32181686f6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.705883] env[62066]: DEBUG nova.network.neutron [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.706356] env[62066]: DEBUG nova.network.neutron [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.811142] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.837759] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156536, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.895080] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.311s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.905090] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.090s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.905569] env[62066]: DEBUG nova.objects.instance [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lazy-loading 'resources' on Instance uuid 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.055010] env[62066]: DEBUG nova.network.neutron [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updated VIF entry in instance network info cache for port f414725f-8af3-4a19-acf5-6cc44ff37ecb. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.055429] env[62066]: DEBUG nova.network.neutron [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updating instance_info_cache with network_info: [{"id": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "address": "fa:16:3e:62:db:6b", "network": {"id": "8090756b-78c4-4068-a3d0-489495be55e0", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-321897902-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.237", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4763071af33b435ba0b7f94b7dcb4de2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4055505f-97ab-400b-969c-43d99b38fd48", "external-id": "nsx-vlan-transportzone-952", "segmentation_id": 952, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf414725f-8a", "ovs_interfaceid": "f414725f-8af3-4a19-acf5-6cc44ff37ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.142212] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.142611] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88620ffd-6b3d-46d1-bce3-14b6a2e016c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.151091] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 957.151091] env[62066]: value = "task-1156537" [ 957.151091] env[62066]: _type = "Task" [ 957.151091] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.160910] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.200520] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c16734-81c8-f269-d1b5-e32181686f6d, 'name': SearchDatastore_Task, 'duration_secs': 0.012428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.201734] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c0fd9ba-c419-4663-bd6c-2f9d0cc816a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.209764] env[62066]: DEBUG oslo_concurrency.lockutils [req-34156954-3757-4b97-b19c-946d5c28e511 req-b2676647-bdf6-4746-82b6-6c9ed006f082 service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.210370] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 957.210370] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52eabbf5-d81b-4a27-1e92-bceb7334e93d" [ 957.210370] env[62066]: _type = "Task" [ 957.210370] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.221356] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52eabbf5-d81b-4a27-1e92-bceb7334e93d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.336725] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751343} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.337026] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 8bc2c334-ff46-44b2-b31a-d85273ec9a45/8bc2c334-ff46-44b2-b31a-d85273ec9a45.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.337263] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.337599] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7ad351d-268b-41b1-af8c-08389af46292 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.359791] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 957.359791] env[62066]: value = "task-1156538" [ 957.359791] env[62066]: _type = "Task" [ 957.359791] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.368983] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.464424] env[62066]: INFO nova.scheduler.client.report [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocation for migration 7806705f-7018-4c75-92eb-f4d8a764f5ed [ 957.558936] env[62066]: DEBUG oslo_concurrency.lockutils [req-1ac56f5d-d68e-4b62-afd3-1e4e469c2d5d req-f6ce4483-2879-47fa-a73a-794caf2d617f service nova] Releasing lock "refresh_cache-16390d89-f2d1-4d6c-be62-b85e61906865" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.583647] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b47df6f-1ba9-496d-a4b6-513b3e47e452 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.587221] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.587456] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.594341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a364402b-585e-420e-acb0-d2c49e8b3a84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.628575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c15304e-46d5-4476-b5f1-de1173b2acf5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.638049] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ad44a3-7204-4221-87ce-4ac0a886637a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.655981] env[62066]: DEBUG nova.compute.provider_tree [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.665732] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156537, 'name': PowerOffVM_Task, 'duration_secs': 0.255477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.666619] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.682631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.682945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.683138] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.683452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.683753] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.686592] env[62066]: INFO nova.compute.manager [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Terminating instance [ 957.688490] env[62066]: DEBUG nova.compute.manager [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.688685] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.689519] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42491be6-fdbd-4d19-a7da-0949350a93e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.697912] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.698042] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0f6f34f-1f24-4863-a211-57670d08909a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.706788] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 957.706788] env[62066]: value = "task-1156539" [ 957.706788] env[62066]: _type = "Task" [ 957.706788] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.716963] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.723150] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52eabbf5-d81b-4a27-1e92-bceb7334e93d, 'name': SearchDatastore_Task, 'duration_secs': 0.011322} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.723459] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.723777] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6e0065b0-da4f-4288-a858-8648572f8148/6e0065b0-da4f-4288-a858-8648572f8148.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 957.724063] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4e1a986-e3b4-4d28-a708-23c7d0028a8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.726824] env[62066]: INFO nova.compute.manager [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Detaching volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 [ 957.735705] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 957.735705] env[62066]: value = "task-1156540" [ 957.735705] env[62066]: _type = "Task" [ 957.735705] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.745353] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.763603] env[62066]: INFO nova.virt.block_device [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Attempting to driver detach volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 from mountpoint /dev/sdb [ 957.763886] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 957.764123] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 957.765031] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6332c865-8fdb-4e0f-bbc6-0f7b760546d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.788566] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5878295-3081-421d-a3a3-fed951c12851 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.797691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922b1940-599f-4301-aaa0-4c0cad40efb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.820034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec994b84-686c-4b38-941a-6c156ac9789e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.837830] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] The volume has not been displaced from its original location: [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 957.843292] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 957.843580] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec5c88c5-b927-4009-85ab-609c338f1628 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.865785] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 957.865785] env[62066]: value = "task-1156541" [ 957.865785] env[62066]: _type = "Task" [ 957.865785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.877422] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.880468] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146505} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.880727] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.881612] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886394e4-8273-45f0-b913-8a8c527f9bd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.904648] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 8bc2c334-ff46-44b2-b31a-d85273ec9a45/8bc2c334-ff46-44b2-b31a-d85273ec9a45.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.905072] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74929a1f-f0d5-4fb5-912f-b7410c7f094f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.927890] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 957.927890] env[62066]: value = "task-1156542" [ 957.927890] env[62066]: _type = "Task" [ 957.927890] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.938586] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156542, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.974384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f86a227e-6084-43d6-815f-8b2ea2c28b1e tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.329s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.095213] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 958.095432] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 958.095594] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 958.163029] env[62066]: DEBUG nova.scheduler.client.report [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.222402] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.251432] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156540, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.379555] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.446711] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156542, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.603023] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Skipping network cache update for instance because it is being deleted. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 958.603023] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Skipping network cache update for instance because it is being deleted. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 958.604751] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 958.604751] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 958.613045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.613473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.613678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.613918] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.614256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.616755] env[62066]: INFO nova.compute.manager [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Terminating instance [ 958.618867] env[62066]: DEBUG nova.compute.manager [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.619070] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.619931] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26af9aec-b901-42e2-bc72-451ac1c8d1ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.630078] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.630078] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36bc8dac-9597-4304-816b-561e42f60545 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.638490] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 958.638490] env[62066]: value = "task-1156543" [ 958.638490] env[62066]: _type = "Task" [ 958.638490] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.652045] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.671315] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.683248] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.683521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.683730] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 958.683911] env[62066]: DEBUG nova.objects.instance [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lazy-loading 'info_cache' on Instance uuid 212dac6a-a291-4ca8-87fb-97ebcca7976c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.696051] env[62066]: INFO nova.scheduler.client.report [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Deleted allocations for instance 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76 [ 958.724286] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.749097] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740703} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.750154] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6e0065b0-da4f-4288-a858-8648572f8148/6e0065b0-da4f-4288-a858-8648572f8148.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.750481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.750805] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4daa90a4-7536-4b65-940f-5606be97b52a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.760579] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 958.760579] env[62066]: value = "task-1156544" [ 958.760579] env[62066]: _type = "Task" [ 958.760579] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.774669] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156544, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.879559] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156541, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.975477] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156542, 'name': ReconfigVM_Task, 'duration_secs': 0.861531} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.975477] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 8bc2c334-ff46-44b2-b31a-d85273ec9a45/8bc2c334-ff46-44b2-b31a-d85273ec9a45.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.975477] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7cbc1314-b19c-4612-9de5-cfe66b4fae9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.975477] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 958.975477] env[62066]: value = "task-1156545" [ 958.975477] env[62066]: _type = "Task" [ 958.975477] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.975477] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156545, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.153651] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156543, 'name': PowerOffVM_Task, 'duration_secs': 0.233379} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.153651] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.153651] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.153651] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57eb65ec-7df6-4ae6-8545-f2e3693c2498 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.204476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e680106-a3cb-4a95-a32b-accbe558468c tempest-ServersNegativeTestJSON-1335163960 tempest-ServersNegativeTestJSON-1335163960-project-member] Lock "0ba970e7-6b21-441a-81f7-2b4e7dfd4d76" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.538s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.219787] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156539, 'name': PowerOffVM_Task, 'duration_secs': 1.124557} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.220462] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.220823] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.221553] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d83e1ca7-ed46-42c0-8fde-a44a884883bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.231059] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.231612] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.231975] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore2] 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.233047] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-787962f7-80a2-4273-ae44-da0ae76ee92e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.243011] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 959.243011] env[62066]: value = "task-1156548" [ 959.243011] env[62066]: _type = "Task" [ 959.243011] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.253520] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156548, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.271936] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156544, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114745} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.271936] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.272733] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946bc231-4b53-4d70-8f77-83f1d168a5c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.298956] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 6e0065b0-da4f-4288-a858-8648572f8148/6e0065b0-da4f-4288-a858-8648572f8148.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.301295] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59171a0f-4bf7-4ae6-9ecf-6af506dc3a5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.319941] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.320145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.320380] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Deleting the datastore file [datastore1] f82c8bf2-24f3-4732-a56d-1bddaef14158 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.320704] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e829a85-8d7a-4f11-9c05-e72b41596cfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.329497] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 959.329497] env[62066]: value = "task-1156550" [ 959.329497] env[62066]: _type = "Task" [ 959.329497] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.329758] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for the task: (returnval){ [ 959.329758] env[62066]: value = "task-1156549" [ 959.329758] env[62066]: _type = "Task" [ 959.329758] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.344121] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.344425] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.380039] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156541, 'name': ReconfigVM_Task, 'duration_secs': 1.021927} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.381799] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 959.389770] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c09a3a5-f87a-4be5-9cca-c14c2e6f78e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.400299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.400553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.409863] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 959.409863] env[62066]: value = "task-1156551" [ 959.409863] env[62066]: _type = "Task" [ 959.409863] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.422037] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156551, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.471627] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156545, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.755485] env[62066]: DEBUG oslo_vmware.api [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156548, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294202} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.755864] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.756115] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.756327] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.756542] env[62066]: INFO nova.compute.manager [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 959.756853] env[62066]: DEBUG oslo.service.loopingcall [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.757101] env[62066]: DEBUG nova.compute.manager [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.757202] env[62066]: DEBUG nova.network.neutron [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 959.849475] env[62066]: DEBUG oslo_vmware.api [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Task: {'id': task-1156549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307421} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.849732] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156550, 'name': ReconfigVM_Task, 'duration_secs': 0.36504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.849967] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.850173] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.850361] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.850542] env[62066]: INFO nova.compute.manager [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Took 2.16 seconds to destroy the instance on the hypervisor. [ 959.850788] env[62066]: DEBUG oslo.service.loopingcall [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.851024] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 6e0065b0-da4f-4288-a858-8648572f8148/6e0065b0-da4f-4288-a858-8648572f8148.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.852278] env[62066]: DEBUG nova.compute.manager [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.852422] env[62066]: DEBUG nova.network.neutron [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 959.854160] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49087e1a-c63f-45e1-8ef5-475bc9e35b7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.862441] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 959.862441] env[62066]: value = "task-1156552" [ 959.862441] env[62066]: _type = "Task" [ 959.862441] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.871570] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156552, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.903468] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.921320] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156551, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.971700] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156545, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.081566] env[62066]: DEBUG nova.compute.manager [req-5f98b6f0-57d9-471d-bff1-a21860ba29f3 req-b5aed17c-d021-4fd6-83e0-c354f2b2dd14 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Received event network-vif-deleted-aef860e5-c337-4d8f-b0ff-e3c92e1ff75f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.081823] env[62066]: INFO nova.compute.manager [req-5f98b6f0-57d9-471d-bff1-a21860ba29f3 req-b5aed17c-d021-4fd6-83e0-c354f2b2dd14 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Neutron deleted interface aef860e5-c337-4d8f-b0ff-e3c92e1ff75f; detaching it from the instance and deleting it from the info cache [ 960.081961] env[62066]: DEBUG nova.network.neutron [req-5f98b6f0-57d9-471d-bff1-a21860ba29f3 req-b5aed17c-d021-4fd6-83e0-c354f2b2dd14 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.271875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.272305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.385026] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156552, 'name': Rename_Task, 'duration_secs': 0.162888} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.385219] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.385568] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0bc10cd-b6a8-4e5c-992c-0fb1c9fc8e9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.396173] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 960.396173] env[62066]: value = "task-1156553" [ 960.396173] env[62066]: _type = "Task" [ 960.396173] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.408262] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.427849] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156551, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.434362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.434838] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.436393] env[62066]: INFO nova.compute.claims [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.473046] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156545, 'name': Rename_Task, 'duration_secs': 1.278147} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.473483] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.473808] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-536a7269-9b2b-4ea6-9d61-4e42f29cd7e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.482116] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 960.482116] env[62066]: value = "task-1156554" [ 960.482116] env[62066]: _type = "Task" [ 960.482116] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.493474] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.510860] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updating instance_info_cache with network_info: [{"id": "59b6e38f-b186-420c-89c8-c860e8310108", "address": "fa:16:3e:5c:97:e8", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59b6e38f-b1", "ovs_interfaceid": "59b6e38f-b186-420c-89c8-c860e8310108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.561112] env[62066]: DEBUG nova.network.neutron [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.585414] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7045299-7a65-44dd-8ef8-3b6f0c2163ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.596790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31aebddb-1ee9-4286-8a6c-b184349c1fae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.633830] env[62066]: DEBUG nova.compute.manager [req-5f98b6f0-57d9-471d-bff1-a21860ba29f3 req-b5aed17c-d021-4fd6-83e0-c354f2b2dd14 service nova] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Detach interface failed, port_id=aef860e5-c337-4d8f-b0ff-e3c92e1ff75f, reason: Instance 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.776753] env[62066]: INFO nova.compute.manager [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Detaching volume 2ac4db90-8817-40a5-a537-07020e6d7eea [ 960.830984] env[62066]: INFO nova.virt.block_device [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Attempting to driver detach volume 2ac4db90-8817-40a5-a537-07020e6d7eea from mountpoint /dev/sdb [ 960.831615] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 960.831858] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251735', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'name': 'volume-2ac4db90-8817-40a5-a537-07020e6d7eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31e50f97-f873-44ad-9923-67923cdb8d3a', 'attached_at': '', 'detached_at': '', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'serial': '2ac4db90-8817-40a5-a537-07020e6d7eea'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 960.832776] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716aa45d-c94c-47cd-8f6f-867c632918b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.865824] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0712b857-6b16-4cde-a622-01deb68a5e8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.876048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f81efe7-4b05-4768-aaed-8174f3cd9ac0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.880248] env[62066]: DEBUG nova.compute.manager [req-8459b16c-137d-402c-b44e-6914b20e98fc req-55a89bf4-78dd-4370-876b-21b39cfa0d0d service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Received event network-vif-deleted-8cbada80-ab13-439c-9a90-eea92446a751 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.880536] env[62066]: INFO nova.compute.manager [req-8459b16c-137d-402c-b44e-6914b20e98fc req-55a89bf4-78dd-4370-876b-21b39cfa0d0d service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Neutron deleted interface 8cbada80-ab13-439c-9a90-eea92446a751; detaching it from the instance and deleting it from the info cache [ 960.880799] env[62066]: DEBUG nova.network.neutron [req-8459b16c-137d-402c-b44e-6914b20e98fc req-55a89bf4-78dd-4370-876b-21b39cfa0d0d service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.921300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cf3a13-07f5-4224-a111-5dbf11c98146 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.930749] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156551, 'name': ReconfigVM_Task, 'duration_secs': 1.196126} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.946588] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 960.950902] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156553, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.953828] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] The volume has not been displaced from its original location: [datastore1] volume-2ac4db90-8817-40a5-a537-07020e6d7eea/volume-2ac4db90-8817-40a5-a537-07020e6d7eea.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 960.958993] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 960.960101] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eac70398-761e-4903-8b93-ebafbe5bca4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.981770] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 960.981770] env[62066]: value = "task-1156555" [ 960.981770] env[62066]: _type = "Task" [ 960.981770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.996722] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156554, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.003679] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156555, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.010435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-212dac6a-a291-4ca8-87fb-97ebcca7976c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.010673] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 961.010886] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011057] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011259] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011471] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011637] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011786] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.011918] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 961.066677] env[62066]: INFO nova.compute.manager [-] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Took 1.31 seconds to deallocate network for instance. [ 961.330793] env[62066]: DEBUG nova.network.neutron [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.388658] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c2391f2-f2ae-4f19-9f66-b85cb0db792b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.402806] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21ecb7b-732c-4eef-af22-145329c2fd96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.425960] env[62066]: DEBUG oslo_vmware.api [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156553, 'name': PowerOnVM_Task, 'duration_secs': 0.548012} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.426614] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.426734] env[62066]: INFO nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Took 8.22 seconds to spawn the instance on the hypervisor. [ 961.427023] env[62066]: DEBUG nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.427910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42488978-156e-4b56-86e2-a25758648a22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.440568] env[62066]: DEBUG nova.compute.manager [req-8459b16c-137d-402c-b44e-6914b20e98fc req-55a89bf4-78dd-4370-876b-21b39cfa0d0d service nova] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Detach interface failed, port_id=8cbada80-ab13-439c-9a90-eea92446a751, reason: Instance f82c8bf2-24f3-4732-a56d-1bddaef14158 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 961.507218] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156555, 'name': ReconfigVM_Task, 'duration_secs': 0.354488} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.507218] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.507218] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82db27ba-c5fa-4d30-bcea-333b930a1c91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.516180] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156554, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.528337] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 961.528337] env[62066]: value = "task-1156556" [ 961.528337] env[62066]: _type = "Task" [ 961.528337] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.543646] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156556, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.573130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.684351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8912794-a9f6-4ee4-a0c3-dfc9c1a4047b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.693404] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f514c420-09cf-4929-bfff-40d7318fb13b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.729850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0630dc33-9a9d-4e4d-93f1-0cefcdcc1371 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.738800] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50659ba5-beeb-4279-8c93-ebb19df1e03c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.753877] env[62066]: DEBUG nova.compute.provider_tree [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.831887] env[62066]: INFO nova.compute.manager [-] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Took 1.98 seconds to deallocate network for instance. [ 961.958270] env[62066]: INFO nova.compute.manager [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Took 13.23 seconds to build instance. [ 961.998688] env[62066]: DEBUG oslo_vmware.api [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156554, 'name': PowerOnVM_Task, 'duration_secs': 1.103281} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.000181] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.000181] env[62066]: INFO nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Took 11.12 seconds to spawn the instance on the hypervisor. [ 962.000181] env[62066]: DEBUG nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.000181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b77c42-4fdc-4236-91d3-f45de86869c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.041582] env[62066]: DEBUG oslo_vmware.api [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156556, 'name': ReconfigVM_Task, 'duration_secs': 0.433836} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.042437] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251735', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'name': 'volume-2ac4db90-8817-40a5-a537-07020e6d7eea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '31e50f97-f873-44ad-9923-67923cdb8d3a', 'attached_at': '', 'detached_at': '', 'volume_id': '2ac4db90-8817-40a5-a537-07020e6d7eea', 'serial': '2ac4db90-8817-40a5-a537-07020e6d7eea'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 962.080135] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.080135] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16722b0e-c353-4581-b069-9a5108b8ab84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.080135] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 962.080135] env[62066]: value = "task-1156557" [ 962.080135] env[62066]: _type = "Task" [ 962.080135] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.080674] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156557, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.259240] env[62066]: DEBUG nova.scheduler.client.report [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.339597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.460432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7fe2f88-f92f-447e-97d7-d1448e592d0c tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.741s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.518769] env[62066]: INFO nova.compute.manager [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Took 19.54 seconds to build instance. [ 962.581888] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 962.582341] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 962.582693] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 962.583754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf9ac62-9d98-4cf6-8a63-0d4f450c4cab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.605411] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85989a5-cc7b-4b05-8c55-a7e3b2ebddd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.616103] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 962.616535] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 962.617476] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6120e579-4cf6-4744-9c38-bdbd84cfa731 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.621868] env[62066]: DEBUG nova.objects.instance [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.631024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.631024] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19654a85-542a-445b-ac12-8f8c78bf0256 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.767023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.767023] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.767692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.195s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.768031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.770541] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.432s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.770902] env[62066]: DEBUG nova.objects.instance [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lazy-loading 'resources' on Instance uuid f82c8bf2-24f3-4732-a56d-1bddaef14158 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.875161] env[62066]: INFO nova.scheduler.client.report [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocations for instance 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f [ 962.875161] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.875161] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.875161] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore1] 00f44ecb-768a-4db3-b229-27bb6f27ede1 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.875161] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f805f761-9703-42f1-8e66-3b12fda55f40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.881027] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 962.881027] env[62066]: value = "task-1156559" [ 962.881027] env[62066]: _type = "Task" [ 962.881027] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.890263] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156559, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.020095] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1cd32cfc-32ac-41ce-94a4-1b1472c81172 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.045s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.276158] env[62066]: DEBUG nova.compute.utils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 963.279957] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.280417] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 963.288296] env[62066]: DEBUG nova.compute.manager [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.288296] env[62066]: DEBUG nova.compute.manager [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.288296] env[62066]: DEBUG oslo_concurrency.lockutils [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.288296] env[62066]: DEBUG oslo_concurrency.lockutils [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.288296] env[62066]: DEBUG nova.network.neutron [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.312943] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0b51b72b-d6ab-4cb9-b215-b913ded67097 tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "2fc3eed6-7af2-4c34-a0d3-0498ae209f5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.699s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.372988] env[62066]: DEBUG nova.policy [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85a0637645c443a9d9e6028053c0450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '350ba3c5676a4dd0a018900e7237a5a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.390074] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156559, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311413} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.390330] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.390507] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.390775] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.478762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd65dbf-918e-49a2-92c5-3ef692db7f56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.488490] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c770bae7-ddcb-40c0-80ce-39d73cfdf2d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.524772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.526315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ee9de5-a937-423d-9eca-47dc74330636 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.535844] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b1f2b4-3698-46fc-bff5-ef10e3eab1c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.553606] env[62066]: DEBUG nova.compute.provider_tree [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.630103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b8f5e48-b457-4c3c-ad72-088f418d1062 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.358s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.634120] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.108s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.634120] env[62066]: DEBUG nova.compute.manager [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.634391] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31c8652-7ebf-45b6-b0d1-7b8e2270b889 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.642645] env[62066]: DEBUG nova.compute.manager [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 963.644689] env[62066]: DEBUG nova.objects.instance [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.664407] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 963.665770] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22590fdd-bc7c-4432-903c-b9c4ad97ff5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.673063] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 963.673272] env[62066]: ERROR oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk due to incomplete transfer. [ 963.673543] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c65005f0-a17d-4af8-bd5b-1abacec40af9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.682727] env[62066]: DEBUG oslo_vmware.rw_handles [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb0827-e3d6-b6d2-5153-4527b02a0755/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 963.682955] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Uploaded image 1b87e52b-3c5d-4619-b6c5-f6075eb5ef9b to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 963.685532] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 963.685813] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3dc2bbcd-4fda-43b1-bb6f-600eb3e24ae3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.693027] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 963.693027] env[62066]: value = "task-1156560" [ 963.693027] env[62066]: _type = "Task" [ 963.693027] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.703096] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156560, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.785107] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 963.898435] env[62066]: INFO nova.virt.block_device [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Booting with volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 at /dev/sdb [ 963.932171] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Successfully created port: 8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.960512] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76607f74-f657-488c-b41d-c875b1704209 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.985258] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b60c4af0-9f15-4807-8c54-546b43b1b28d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.030789] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9479c2c-f2f1-4305-99a5-9b3900aa4f59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.044780] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140bc841-1898-4c60-bab4-4166db7c2b32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.061549] env[62066]: DEBUG nova.scheduler.client.report [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.089718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715fa2a1-51a8-40aa-a433-8d580e945b39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.098179] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4730f40-1a25-4896-ab11-95a565358e80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.116965] env[62066]: DEBUG nova.virt.block_device [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating existing volume attachment record: 8c6c8bff-b568-4443-a7fc-45f325f5cff0 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 964.151106] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.151509] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fedf95b-7940-4b15-ada8-b79dc92d4507 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.162776] env[62066]: DEBUG oslo_vmware.api [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 964.162776] env[62066]: value = "task-1156561" [ 964.162776] env[62066]: _type = "Task" [ 964.162776] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.173975] env[62066]: DEBUG oslo_vmware.api [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156561, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.204234] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156560, 'name': Destroy_Task, 'duration_secs': 0.449788} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.204613] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Destroyed the VM [ 964.204861] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 964.205142] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-52dd7d6d-f815-4f60-9833-a4fc9ae30a26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.217830] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 964.217830] env[62066]: value = "task-1156562" [ 964.217830] env[62066]: _type = "Task" [ 964.217830] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.232753] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156562, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.233838] env[62066]: DEBUG nova.network.neutron [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.234213] env[62066]: DEBUG nova.network.neutron [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.451385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.451664] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.451858] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.452085] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.452358] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.454946] env[62066]: INFO nova.compute.manager [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Terminating instance [ 964.457207] env[62066]: DEBUG nova.compute.manager [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.457299] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.458286] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12370b80-0082-489c-a369-87f7085a5071 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.468136] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.468455] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-507075a3-3ef1-40bf-8a6a-8673c2ea411c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.476580] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 964.476580] env[62066]: value = "task-1156563" [ 964.476580] env[62066]: _type = "Task" [ 964.476580] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.487043] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.580405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.580405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.580405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.614571] env[62066]: INFO nova.scheduler.client.report [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Deleted allocations for instance f82c8bf2-24f3-4732-a56d-1bddaef14158 [ 964.673677] env[62066]: DEBUG oslo_vmware.api [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156561, 'name': PowerOffVM_Task, 'duration_secs': 0.238909} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.673901] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.674109] env[62066]: DEBUG nova.compute.manager [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.674965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55065416-f0e0-4ea8-9842-0d10928265d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.730229] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156562, 'name': RemoveSnapshot_Task, 'duration_secs': 0.429397} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.730508] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 964.730789] env[62066]: DEBUG nova.compute.manager [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.731810] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a391e6-3f6d-48be-9e0a-49de73810866 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.736471] env[62066]: DEBUG oslo_concurrency.lockutils [req-f682ea40-d681-47ad-8fca-a703fdea5114 req-2e338ed9-c6db-41a0-b118-3b702d3982ae service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.799708] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 964.838102] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.838414] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.838581] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.838769] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.839036] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.839296] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.839535] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.839705] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.839881] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.840061] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.840248] env[62066]: DEBUG nova.virt.hardware [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.841711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90408462-d30e-4a6a-a42a-9da0ea2794e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.850766] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbea2c4b-0602-4735-9007-8cc278c40e9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.986949] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156563, 'name': PowerOffVM_Task, 'duration_secs': 0.225247} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.989784] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.989784] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.989784] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6faced3f-c279-4f74-9212-a520c726fa0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.080856] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.080856] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.080856] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] 8bc2c334-ff46-44b2-b31a-d85273ec9a45 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.080856] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d06f83df-1271-45e0-9b21-525409c7151f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.086077] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 965.094464] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 965.094464] env[62066]: value = "task-1156565" [ 965.094464] env[62066]: _type = "Task" [ 965.094464] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.110401] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.125121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63052de9-10dc-416b-b3b3-bc23e355e901 tempest-ServersV294TestFqdnHostnames-1049263718 tempest-ServersV294TestFqdnHostnames-1049263718-project-member] Lock "f82c8bf2-24f3-4732-a56d-1bddaef14158" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.442s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.200747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3f664c1b-c8e6-4acd-9720-44cd2ce58fe8 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.567s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.247790] env[62066]: INFO nova.compute.manager [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Shelve offloading [ 965.256172] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.256680] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09e36415-4a37-4317-802d-7ebbaf400701 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.267766] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 965.267766] env[62066]: value = "task-1156566" [ 965.267766] env[62066]: _type = "Task" [ 965.267766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.280852] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 965.281229] env[62066]: DEBUG nova.compute.manager [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.285222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f914e5-d12e-4bac-b22d-5ad19164e5dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.294306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.294607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.294684] env[62066]: DEBUG nova.network.neutron [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.606982] env[62066]: DEBUG oslo_vmware.api [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158785} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.607565] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.607866] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.609239] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.609239] env[62066]: INFO nova.compute.manager [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Took 1.15 seconds to destroy the instance on the hypervisor. [ 965.609239] env[62066]: DEBUG oslo.service.loopingcall [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.609239] env[62066]: DEBUG nova.compute.manager [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.609239] env[62066]: DEBUG nova.network.neutron [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 965.615103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.615103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.615279] env[62066]: INFO nova.compute.claims [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.669268] env[62066]: DEBUG nova.compute.manager [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Received event network-vif-plugged-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.669501] env[62066]: DEBUG oslo_concurrency.lockutils [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] Acquiring lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.669784] env[62066]: DEBUG oslo_concurrency.lockutils [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.669983] env[62066]: DEBUG oslo_concurrency.lockutils [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.670641] env[62066]: DEBUG nova.compute.manager [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] No waiting events found dispatching network-vif-plugged-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.670871] env[62066]: WARNING nova.compute.manager [req-aee672c7-cd48-48c7-ba38-60856c3bbf61 req-3cb9b090-202f-47ff-8163-26287c2d8c2e service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Received unexpected event network-vif-plugged-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 for instance with vm_state building and task_state spawning. [ 965.865658] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Successfully updated port: 8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.150474] env[62066]: DEBUG nova.objects.instance [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.251304] env[62066]: DEBUG nova.compute.manager [req-ccbc5fc7-481c-4e1e-b581-f0f2a4da6622 req-52f7b013-cdf3-4bd2-b4bb-42c21c2e666e service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Received event network-vif-deleted-ef31c8ff-7052-4009-9b1c-64674449bae4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.251304] env[62066]: INFO nova.compute.manager [req-ccbc5fc7-481c-4e1e-b581-f0f2a4da6622 req-52f7b013-cdf3-4bd2-b4bb-42c21c2e666e service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Neutron deleted interface ef31c8ff-7052-4009-9b1c-64674449bae4; detaching it from the instance and deleting it from the info cache [ 966.251304] env[62066]: DEBUG nova.network.neutron [req-ccbc5fc7-481c-4e1e-b581-f0f2a4da6622 req-52f7b013-cdf3-4bd2-b4bb-42c21c2e666e service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.291668] env[62066]: DEBUG nova.network.neutron [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updating instance_info_cache with network_info: [{"id": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "address": "fa:16:3e:6a:8b:8e", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14b4402d-b3", "ovs_interfaceid": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.301768] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.302064] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.302297] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.302570] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.302811] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.302998] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.303241] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.304617] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.304617] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.304617] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.304617] env[62066]: DEBUG nova.virt.hardware [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.306659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257c94e6-bcfa-4b4b-8f4a-8260538ee820 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.315995] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d84f79-b1a3-4361-880d-4087fbc4ce5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.333096] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:82:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0fae426-1b77-435b-98a6-72d487ff589f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.343169] env[62066]: DEBUG oslo.service.loopingcall [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.344165] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 966.344870] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f9c1cb9-2cb8-4f27-b888-bc0234d92fce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.368530] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.368530] env[62066]: value = "task-1156567" [ 966.368530] env[62066]: _type = "Task" [ 966.368530] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.372521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.372677] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.372976] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 966.379917] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156567, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.657048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.657048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.657048] env[62066]: DEBUG nova.network.neutron [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 966.657338] env[62066]: DEBUG nova.objects.instance [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'info_cache' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.720554] env[62066]: DEBUG nova.network.neutron [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.754537] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f72c1eed-3bc3-4a80-ab3c-1b22006ee3d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.769650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7e92e5-732e-40cd-af67-d90a7398db8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.797408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.810266] env[62066]: DEBUG nova.compute.manager [req-ccbc5fc7-481c-4e1e-b581-f0f2a4da6622 req-52f7b013-cdf3-4bd2-b4bb-42c21c2e666e service nova] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Detach interface failed, port_id=ef31c8ff-7052-4009-9b1c-64674449bae4, reason: Instance 8bc2c334-ff46-44b2-b31a-d85273ec9a45 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.862265] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf30d614-51f2-4298-92ce-9cb24a142a58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.877639] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd7f1c8-f545-4c55-a7e9-8fc312ef2ea7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.893737] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156567, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.920096] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d25269a-9f28-49ce-9e87-b07ae2abbcde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.931066] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6d2162-b100-40a7-961c-8f15a5d6731c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.949320] env[62066]: DEBUG nova.compute.provider_tree [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.950673] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 967.163066] env[62066]: DEBUG nova.objects.base [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Object Instance<31e50f97-f873-44ad-9923-67923cdb8d3a> lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 967.229108] env[62066]: INFO nova.compute.manager [-] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Took 1.62 seconds to deallocate network for instance. [ 967.299704] env[62066]: DEBUG nova.network.neutron [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Updating instance_info_cache with network_info: [{"id": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "address": "fa:16:3e:5c:41:16", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c87851b-47", "ovs_interfaceid": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.382746] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156567, 'name': CreateVM_Task, 'duration_secs': 0.628067} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.382949] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.384477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.384477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.385052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.385052] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e054708-6fe5-4b2c-960e-62c2b8f902aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.392035] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 967.392035] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5206babc-fee1-35a4-bb59-63399a96a5a1" [ 967.392035] env[62066]: _type = "Task" [ 967.392035] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.400748] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5206babc-fee1-35a4-bb59-63399a96a5a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.454612] env[62066]: DEBUG nova.scheduler.client.report [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.647326] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.648508] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9a19d6-4118-40b0-9985-1735556306d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.659143] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.659440] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28bb9f98-c24d-4544-a574-83171cbca1e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.697081] env[62066]: DEBUG nova.compute.manager [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Received event network-changed-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.697332] env[62066]: DEBUG nova.compute.manager [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Refreshing instance network info cache due to event network-changed-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.697546] env[62066]: DEBUG oslo_concurrency.lockutils [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] Acquiring lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.737322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.753168] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.753168] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.753168] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore2] 3e7a2dd5-f01e-4edd-884f-7d614f260181 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.753168] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac44cebf-26b1-41ab-b894-d7adef359ff4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.761078] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 967.761078] env[62066]: value = "task-1156569" [ 967.761078] env[62066]: _type = "Task" [ 967.761078] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.779964] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.802752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.803152] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Instance network_info: |[{"id": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "address": "fa:16:3e:5c:41:16", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c87851b-47", "ovs_interfaceid": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.803596] env[62066]: DEBUG oslo_concurrency.lockutils [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] Acquired lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.803812] env[62066]: DEBUG nova.network.neutron [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Refreshing network info cache for port 8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 967.805353] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:41:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be5c038c-29e5-43c9-91ab-9eb3094b5337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c87851b-47ba-4dcf-abe5-ab414ef3d4b5', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.818972] env[62066]: DEBUG oslo.service.loopingcall [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.820266] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.820549] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe26026b-c64a-40c7-84f1-f54fe6a04a52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.848390] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.848390] env[62066]: value = "task-1156570" [ 967.848390] env[62066]: _type = "Task" [ 967.848390] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.860196] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156570, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.911454] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5206babc-fee1-35a4-bb59-63399a96a5a1, 'name': SearchDatastore_Task, 'duration_secs': 0.01006} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.911454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.911454] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.911454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.911454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.912334] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.912334] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf4d5908-0236-4323-a275-88423df7790f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.922950] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.923231] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.924096] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ad425e-95a1-4235-8239-db13e7823920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.932998] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 967.932998] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525229d2-39d6-9a3b-8de0-8536c54db570" [ 967.932998] env[62066]: _type = "Task" [ 967.932998] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.943951] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525229d2-39d6-9a3b-8de0-8536c54db570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.961554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.961554] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 967.965497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.228s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.966078] env[62066]: DEBUG nova.objects.instance [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'resources' on Instance uuid 8bc2c334-ff46-44b2-b31a-d85273ec9a45 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.031039] env[62066]: DEBUG nova.network.neutron [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [{"id": "6244eb3b-c073-494b-b711-f8712351da75", "address": "fa:16:3e:bf:23:d2", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6244eb3b-c0", "ovs_interfaceid": "6244eb3b-c073-494b-b711-f8712351da75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.274069] env[62066]: DEBUG oslo_vmware.api [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170984} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.274470] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.274702] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.274871] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.300585] env[62066]: DEBUG nova.compute.manager [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received event network-vif-unplugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.300837] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.302546] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.302946] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.303043] env[62066]: DEBUG nova.compute.manager [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] No waiting events found dispatching network-vif-unplugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 968.303232] env[62066]: WARNING nova.compute.manager [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received unexpected event network-vif-unplugged-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 for instance with vm_state shelved and task_state shelving_offloading. [ 968.303440] env[62066]: DEBUG nova.compute.manager [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Received event network-changed-14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.303604] env[62066]: DEBUG nova.compute.manager [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Refreshing instance network info cache due to event network-changed-14b4402d-b3d0-42ef-8c2c-c02c225d8b24. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.303782] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Acquiring lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.303923] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Acquired lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.304145] env[62066]: DEBUG nova.network.neutron [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Refreshing network info cache for port 14b4402d-b3d0-42ef-8c2c-c02c225d8b24 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.362705] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156570, 'name': CreateVM_Task, 'duration_secs': 0.404392} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.362705] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.364384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.364601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.364815] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.365351] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ae5cfb7-3351-4df0-9745-fa89ccd29a4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.369864] env[62066]: INFO nova.scheduler.client.report [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 3e7a2dd5-f01e-4edd-884f-7d614f260181 [ 968.375766] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 968.375766] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52056cd2-a969-c64b-86ef-c4eb9ca134ed" [ 968.375766] env[62066]: _type = "Task" [ 968.375766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.385612] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52056cd2-a969-c64b-86ef-c4eb9ca134ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.450533] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525229d2-39d6-9a3b-8de0-8536c54db570, 'name': SearchDatastore_Task, 'duration_secs': 0.015053} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.450533] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-212e00c4-3267-4765-9428-fe14c83d430d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.456882] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 968.456882] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5274404a-d98f-0084-27c4-aad3473f4927" [ 968.456882] env[62066]: _type = "Task" [ 968.456882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.472976] env[62066]: DEBUG nova.compute.utils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 968.474428] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5274404a-d98f-0084-27c4-aad3473f4927, 'name': SearchDatastore_Task, 'duration_secs': 0.011599} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.475025] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 968.475733] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 968.476947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.477224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 968.477880] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f8d1c62-7163-448f-9032-afe37a43be7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.490495] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 968.490495] env[62066]: value = "task-1156571" [ 968.490495] env[62066]: _type = "Task" [ 968.490495] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.500162] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.535588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "refresh_cache-31e50f97-f873-44ad-9923-67923cdb8d3a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.550326] env[62066]: DEBUG nova.policy [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 968.688175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22af2ddb-150c-4c7c-8f38-c57777707dc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.692173] env[62066]: DEBUG nova.network.neutron [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Updated VIF entry in instance network info cache for port 8c87851b-47ba-4dcf-abe5-ab414ef3d4b5. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 968.693168] env[62066]: DEBUG nova.network.neutron [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Updating instance_info_cache with network_info: [{"id": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "address": "fa:16:3e:5c:41:16", "network": {"id": "70556395-1275-47a0-8234-3c1df611aa19", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006745375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "350ba3c5676a4dd0a018900e7237a5a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be5c038c-29e5-43c9-91ab-9eb3094b5337", "external-id": "nsx-vlan-transportzone-511", "segmentation_id": 511, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c87851b-47", "ovs_interfaceid": "8c87851b-47ba-4dcf-abe5-ab414ef3d4b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.704391] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248c3490-679a-49ca-85c6-4719389e6ddf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.738808] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d8273a-8c53-4556-8cff-4cecca2c236e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.748280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bbd837-6515-4823-a35a-16f4d365f434 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.765841] env[62066]: DEBUG nova.compute.provider_tree [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.875771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.892149] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52056cd2-a969-c64b-86ef-c4eb9ca134ed, 'name': SearchDatastore_Task, 'duration_secs': 0.015529} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.892558] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.892854] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.893213] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.893410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.893682] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.894041] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9c88ec0-ea88-49b2-9ad5-0839debea522 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.907737] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.908024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.908904] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea035923-919c-4565-9f95-8405484cc14e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.918387] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 968.918387] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5221e4c0-ac53-eba0-e7ac-1997c695ec24" [ 968.918387] env[62066]: _type = "Task" [ 968.918387] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.937508] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5221e4c0-ac53-eba0-e7ac-1997c695ec24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.975946] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 969.004857] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156571, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.044533] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.045149] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-090f52f4-7eb4-4b36-a61d-3daeb06bc49a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.054262] env[62066]: DEBUG oslo_vmware.api [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 969.054262] env[62066]: value = "task-1156572" [ 969.054262] env[62066]: _type = "Task" [ 969.054262] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.064672] env[62066]: DEBUG oslo_vmware.api [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.096156] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Successfully created port: 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.175103] env[62066]: DEBUG nova.network.neutron [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updated VIF entry in instance network info cache for port 14b4402d-b3d0-42ef-8c2c-c02c225d8b24. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.175515] env[62066]: DEBUG nova.network.neutron [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updating instance_info_cache with network_info: [{"id": "14b4402d-b3d0-42ef-8c2c-c02c225d8b24", "address": "fa:16:3e:6a:8b:8e", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": null, "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap14b4402d-b3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.196023] env[62066]: DEBUG oslo_concurrency.lockutils [req-da687674-3dc8-4d8b-ac59-3fae44d86760 req-0a03f6c5-8a72-478e-b965-0231e66f6feb service nova] Releasing lock "refresh_cache-10a2f245-93fd-4f6a-8c2f-897fd82269d3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.272023] env[62066]: DEBUG nova.scheduler.client.report [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.431508] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5221e4c0-ac53-eba0-e7ac-1997c695ec24, 'name': SearchDatastore_Task, 'duration_secs': 0.02267} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.432494] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-235478db-a7ce-494b-83f6-655e013e69e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.439121] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 969.439121] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52897543-9e7e-3bad-7d98-bd751faa34cc" [ 969.439121] env[62066]: _type = "Task" [ 969.439121] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.448786] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52897543-9e7e-3bad-7d98-bd751faa34cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.502805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "845748f7-d87b-4141-9932-235acd3d764c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.503086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.508962] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56138} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.512405] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.512405] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.512405] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8394d943-c200-4891-a918-e269e30d3c20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.521713] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 969.521713] env[62066]: value = "task-1156573" [ 969.521713] env[62066]: _type = "Task" [ 969.521713] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.533487] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.564025] env[62066]: DEBUG oslo_vmware.api [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156572, 'name': PowerOnVM_Task, 'duration_secs': 0.452652} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.564362] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 969.564609] env[62066]: DEBUG nova.compute.manager [None req-03c34c91-ac6b-4809-ae4d-4bd2a5ff4600 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.565412] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdfda52-620f-49ea-9d18-f801b4c52a58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.679433] env[62066]: DEBUG oslo_concurrency.lockutils [req-7baa52f1-e4fb-472e-af25-9591f2eed227 req-3b79d0b9-5e48-493e-9e36-491644b36785 service nova] Releasing lock "refresh_cache-3e7a2dd5-f01e-4edd-884f-7d614f260181" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.775262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.781924] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.905s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.781924] env[62066]: DEBUG nova.objects.instance [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'resources' on Instance uuid 3e7a2dd5-f01e-4edd-884f-7d614f260181 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.812679] env[62066]: INFO nova.scheduler.client.report [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 8bc2c334-ff46-44b2-b31a-d85273ec9a45 [ 969.949756] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52897543-9e7e-3bad-7d98-bd751faa34cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010537} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.949888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.950049] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 10a2f245-93fd-4f6a-8c2f-897fd82269d3/10a2f245-93fd-4f6a-8c2f-897fd82269d3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.951043] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdc58986-9518-4eee-a1cf-2de6208f1db8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.956955] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 969.956955] env[62066]: value = "task-1156574" [ 969.956955] env[62066]: _type = "Task" [ 969.956955] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.964737] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156574, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.989956] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 970.010425] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.015700] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.015988] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.016170] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.016359] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.016509] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.016662] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.016870] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.017041] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.017220] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.017385] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.017569] env[62066]: DEBUG nova.virt.hardware [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.018050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.018888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca4b192-aa44-459a-8945-2a76058bd5f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.028673] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6cd27d-d0b9-434a-97a6-4f4b226e2b31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.039057] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076687} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.039057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.039057] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af98fea2-f29f-4e63-a9ca-53d9fc257478 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.068532] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.069564] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba575b42-775a-44ed-bf35-bbab22b604d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.094119] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 970.094119] env[62066]: value = "task-1156575" [ 970.094119] env[62066]: _type = "Task" [ 970.094119] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.104019] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.286166] env[62066]: DEBUG nova.objects.instance [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'numa_topology' on Instance uuid 3e7a2dd5-f01e-4edd-884f-7d614f260181 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.322711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2b96174-9d21-4755-b91b-51ac4fe984d2 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "8bc2c334-ff46-44b2-b31a-d85273ec9a45" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.871s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.470372] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156574, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.549166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.595774] env[62066]: DEBUG nova.compute.manager [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-vif-plugged-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.595990] env[62066]: DEBUG oslo_concurrency.lockutils [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.596500] env[62066]: DEBUG oslo_concurrency.lockutils [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.596688] env[62066]: DEBUG oslo_concurrency.lockutils [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.596855] env[62066]: DEBUG nova.compute.manager [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] No waiting events found dispatching network-vif-plugged-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.597030] env[62066]: WARNING nova.compute.manager [req-5acccc70-264f-4947-bf0a-1b8c74d53063 req-9d6d7fb1-f698-49f4-8d50-f63ad0e28d25 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received unexpected event network-vif-plugged-567b53ff-9743-4fa0-b6ed-37cab139833f for instance with vm_state building and task_state spawning. [ 970.607474] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156575, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.684986] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Successfully updated port: 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.792949] env[62066]: DEBUG nova.objects.base [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Object Instance<3e7a2dd5-f01e-4edd-884f-7d614f260181> lazy-loaded attributes: resources,numa_topology {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 970.965489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8d05e5-66ea-4885-8e36-3e3aac246ead {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.973866] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156574, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656563} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.977321] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 10a2f245-93fd-4f6a-8c2f-897fd82269d3/10a2f245-93fd-4f6a-8c2f-897fd82269d3.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.977321] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.977321] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7408b462-9a3f-479c-8f78-f7aae0ddbe3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.978992] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a2fb1f-3640-494f-8851-68d732525036 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.014942] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e52983-95ce-43d7-8662-52283b503732 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.019116] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 971.019116] env[62066]: value = "task-1156576" [ 971.019116] env[62066]: _type = "Task" [ 971.019116] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.029174] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4e6310-3b1d-4ee6-8bfc-9771a0ad21e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.038088] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156576, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.047641] env[62066]: DEBUG nova.compute.provider_tree [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.108376] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156575, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.187336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.187545] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.187757] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.528364] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068562} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.528660] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.529466] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1ca610-d0aa-4948-8f48-d86e4a9a94b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.551833] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 10a2f245-93fd-4f6a-8c2f-897fd82269d3/10a2f245-93fd-4f6a-8c2f-897fd82269d3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.552723] env[62066]: DEBUG nova.scheduler.client.report [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.555920] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28032eea-ffe8-4f91-8e49-2aa0d65e7b43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.576019] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 971.576019] env[62066]: value = "task-1156577" [ 971.576019] env[62066]: _type = "Task" [ 971.576019] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.585769] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.607212] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156575, 'name': ReconfigVM_Task, 'duration_secs': 1.298554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.607511] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 00f44ecb-768a-4db3-b229-27bb6f27ede1/00f44ecb-768a-4db3-b229-27bb6f27ede1.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.608874] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'encryption_format': None, 'encrypted': False, 'device_name': '/dev/sda', 'device_type': 'disk', 'encryption_secret_uuid': None, 'boot_index': 0, 'size': 0, 'guest_format': None, 'encryption_options': None, 'image_id': 'fc5145ed-66bc-4490-b8ac-7ca0de814dc0'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'device_type': None, 'boot_index': None, 'attachment_id': '8c6c8bff-b568-4443-a7fc-45f325f5cff0', 'mount_device': '/dev/sdb', 'guest_format': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'}, 'volume_type': None}], 'swap': None} {{(pid=62066) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 971.609095] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 971.609296] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 971.610082] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd37a14-45a0-490c-8f17-f70321fe8b8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.625881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3d9a78-755c-4844-8f18-ae551c517f57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.650848] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.651169] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87cab5df-fc35-4031-a49a-59e7f90a8e5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.670065] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 971.670065] env[62066]: value = "task-1156578" [ 971.670065] env[62066]: _type = "Task" [ 971.670065] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.678697] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156578, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.726161] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 971.888169] env[62066]: DEBUG nova.network.neutron [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.936169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.936418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.071022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.290s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.077023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.525s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.077023] env[62066]: INFO nova.compute.claims [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.089631] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156577, 'name': ReconfigVM_Task, 'duration_secs': 0.4586} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.089932] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 10a2f245-93fd-4f6a-8c2f-897fd82269d3/10a2f245-93fd-4f6a-8c2f-897fd82269d3.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.090758] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6440c906-379e-490a-8ab4-7a26f6b27e3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.098792] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 972.098792] env[62066]: value = "task-1156579" [ 972.098792] env[62066]: _type = "Task" [ 972.098792] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.110504] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156579, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.180477] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156578, 'name': ReconfigVM_Task, 'duration_secs': 0.401474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.180739] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.185360] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f51a39cb-8c2b-4c88-aa51-5644da6aeadb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.200301] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 972.200301] env[62066]: value = "task-1156580" [ 972.200301] env[62066]: _type = "Task" [ 972.200301] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.208607] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156580, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.391961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.391961] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Instance network_info: |[{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 972.392374] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:8c:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '65497291-07f3-434c-bd42-657a0cb03365', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '567b53ff-9743-4fa0-b6ed-37cab139833f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.400098] env[62066]: DEBUG oslo.service.loopingcall [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.400375] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.400617] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18183196-b9fd-47f9-9632-41f8d1dea602 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.420652] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.420652] env[62066]: value = "task-1156581" [ 972.420652] env[62066]: _type = "Task" [ 972.420652] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.428201] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156581, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.438862] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.585623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59b6316a-efc0-4c17-8956-6ebac3c09f41 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.454s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.586563] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.569s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.586838] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.587106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.587325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.589706] env[62066]: INFO nova.compute.manager [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Terminating instance [ 972.592515] env[62066]: DEBUG nova.compute.manager [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.592709] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.593027] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9edc02c9-2fe7-4f06-97fa-bde91df9fa4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.606704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22f9b5e-81d7-49ee-af14-a161fdf8c0f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.623237] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156579, 'name': Rename_Task, 'duration_secs': 0.226436} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.624389] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.625555] env[62066]: DEBUG nova.compute.manager [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.625744] env[62066]: DEBUG nova.compute.manager [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing instance network info cache due to event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 972.626939] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.626939] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.626939] env[62066]: DEBUG nova.network.neutron [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.640307] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a0b813d-6bba-4b57-a1ff-92a032549322 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.643352] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3e7a2dd5-f01e-4edd-884f-7d614f260181 could not be found. [ 972.643594] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.643778] env[62066]: INFO nova.compute.manager [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Took 0.05 seconds to destroy the instance on the hypervisor. [ 972.644027] env[62066]: DEBUG oslo.service.loopingcall [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.649152] env[62066]: DEBUG nova.compute.manager [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.649152] env[62066]: DEBUG nova.network.neutron [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.655599] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 972.655599] env[62066]: value = "task-1156582" [ 972.655599] env[62066]: _type = "Task" [ 972.655599] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.668670] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.711496] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156580, 'name': ReconfigVM_Task, 'duration_secs': 0.185482} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.711859] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 972.712537] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7899d4c-67ea-4fc5-935a-7ef0fd8f4e80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.719395] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 972.719395] env[62066]: value = "task-1156583" [ 972.719395] env[62066]: _type = "Task" [ 972.719395] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.728314] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156583, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.927560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.927945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.934739] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156581, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.971238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.003825] env[62066]: DEBUG nova.network.neutron [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updated VIF entry in instance network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 973.004344] env[62066]: DEBUG nova.network.neutron [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.166730] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156582, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.226878] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9028370e-a623-4766-a7a7-f8f04dde68d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.232261] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156583, 'name': Rename_Task, 'duration_secs': 0.182438} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.232875] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 973.233143] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d8fdf2f-110f-407a-85c2-ce415e76529a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.238231] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932b8fe8-f6a8-4c32-8aa5-46e5711c34a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.242816] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 973.242816] env[62066]: value = "task-1156584" [ 973.242816] env[62066]: _type = "Task" [ 973.242816] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.276067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e4a305-fa23-49e6-95e6-663796ef7a09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.282118] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156584, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.287042] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f3468c-b7b5-48dd-9c07-5e0bb195d361 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.302061] env[62066]: DEBUG nova.compute.provider_tree [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.436661] env[62066]: DEBUG nova.compute.utils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 973.436661] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156581, 'name': CreateVM_Task, 'duration_secs': 0.536564} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.436661] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 973.436661] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.436661] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.436995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.437130] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8a34afd-e2d6-4b20-99a6-bc06bd56b888 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.442446] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 973.442446] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]522bee24-da8f-c0ba-8de2-a62cb38ea987" [ 973.442446] env[62066]: _type = "Task" [ 973.442446] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.451837] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522bee24-da8f-c0ba-8de2-a62cb38ea987, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.507769] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdcd89df-3834-4726-9099-0e13e5be503c req-5dfd981d-5c00-4f77-b0a8-84f30193b000 service nova] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.596240] env[62066]: DEBUG nova.network.neutron [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.665860] env[62066]: DEBUG oslo_vmware.api [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156582, 'name': PowerOnVM_Task, 'duration_secs': 0.765164} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.666215] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.666430] env[62066]: INFO nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Took 8.87 seconds to spawn the instance on the hypervisor. [ 973.666615] env[62066]: DEBUG nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.667398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f32024-e4bb-4d62-a3b1-7e17f10e1948 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.752519] env[62066]: DEBUG oslo_vmware.api [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156584, 'name': PowerOnVM_Task, 'duration_secs': 0.507673} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.752816] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.753048] env[62066]: DEBUG nova.compute.manager [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.753914] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d862e5-9ba6-41d8-a7bd-9bb855c201f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.804032] env[62066]: DEBUG nova.scheduler.client.report [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.936624] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.954901] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]522bee24-da8f-c0ba-8de2-a62cb38ea987, 'name': SearchDatastore_Task, 'duration_secs': 0.010914} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.955228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.955460] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 973.955700] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.955853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.956053] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 973.956311] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cf7e133-614f-4c46-bae6-95cc5c3195d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.964737] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 973.964931] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 973.965655] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ecb689-8923-4eb5-9a66-514f8f5932da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.970985] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 973.970985] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd4505-10c0-f6a7-bbf6-e96012f3e45c" [ 973.970985] env[62066]: _type = "Task" [ 973.970985] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.978388] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd4505-10c0-f6a7-bbf6-e96012f3e45c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.097912] env[62066]: INFO nova.compute.manager [-] [instance: 3e7a2dd5-f01e-4edd-884f-7d614f260181] Took 1.45 seconds to deallocate network for instance. [ 974.186171] env[62066]: INFO nova.compute.manager [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Took 13.78 seconds to build instance. [ 974.269217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.309435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.310121] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.312730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.342s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.314732] env[62066]: INFO nova.compute.claims [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.481888] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52cd4505-10c0-f6a7-bbf6-e96012f3e45c, 'name': SearchDatastore_Task, 'duration_secs': 0.009701} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.482785] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e1efc4e-efa8-4c93-a1bd-59f7d5674546 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.488499] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 974.488499] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52be0f69-2f2c-15a9-953a-f25171553b7c" [ 974.488499] env[62066]: _type = "Task" [ 974.488499] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.501774] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52be0f69-2f2c-15a9-953a-f25171553b7c, 'name': SearchDatastore_Task, 'duration_secs': 0.010329} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.501774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.501774] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 045c2efa-e2fe-4d51-b267-e2b380fcf1d2/045c2efa-e2fe-4d51-b267-e2b380fcf1d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.501964] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb3c76e9-7ca1-496e-8918-4a3a26da17c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.508093] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 974.508093] env[62066]: value = "task-1156585" [ 974.508093] env[62066]: _type = "Task" [ 974.508093] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.516376] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.688533] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27537064-218e-4d1b-a9e0-af37ceb6295b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.288s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.819377] env[62066]: DEBUG nova.compute.utils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 974.824366] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.824597] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 974.895014] env[62066]: DEBUG nova.policy [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfbb0d60fede4d709342e513540c72b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e6417ca1e084cf5889618dee4e86718', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.998093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.998489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.998489] env[62066]: INFO nova.compute.manager [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Attaching volume 9efbca39-3380-44d5-a311-1c9255ac3a78 to /dev/sdb [ 975.016740] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156585, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.033928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03b53c7-d260-4137-a6d6-cb8f04177004 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.041781] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e0e84f-8023-4e96-a2ad-6ea7e6badf76 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.054742] env[62066]: DEBUG nova.virt.block_device [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating existing volume attachment record: af7c2c26-e54e-45ec-8dd9-fd0f4b4090a9 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 975.134812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9c807669-edb1-48c6-a659-7d5ec25af4ef tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "3e7a2dd5-f01e-4edd-884f-7d614f260181" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.547s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.166285] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Successfully created port: bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.325372] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.349035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.349359] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.427428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.427428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.427606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.427748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.427935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.431335] env[62066]: INFO nova.compute.manager [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Terminating instance [ 975.437177] env[62066]: DEBUG nova.compute.manager [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.438022] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.438510] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02768796-1f8c-4540-b718-3a4bb4fca8d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.447626] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.450300] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-287e71d4-4e23-461b-8380-80f082e4b659 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.456605] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 975.456605] env[62066]: value = "task-1156589" [ 975.456605] env[62066]: _type = "Task" [ 975.456605] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.466810] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.507570] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93a3aec-c2e0-445f-9c60-7c8a463d5f05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.520325] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efff1782-7755-498e-956b-c629e8305e5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.523932] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511348} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.524198] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 045c2efa-e2fe-4d51-b267-e2b380fcf1d2/045c2efa-e2fe-4d51-b267-e2b380fcf1d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 975.524436] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.525199] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4149b8b2-4445-4262-a0a3-0e5120e8e076 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.554389] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aebb736-7c3f-485d-bad6-5cd7edcb4dd8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.562181] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 975.562181] env[62066]: value = "task-1156590" [ 975.562181] env[62066]: _type = "Task" [ 975.562181] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.565278] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb6fbf4-b355-4323-872e-96e96cec7c30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.585021] env[62066]: DEBUG nova.compute.provider_tree [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.586430] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.855519] env[62066]: INFO nova.compute.manager [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Detaching volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 [ 975.887743] env[62066]: INFO nova.virt.block_device [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Attempting to driver detach volume 0b5166a2-4e24-4bee-8c6e-0591e1e80700 from mountpoint /dev/sdb [ 975.888067] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 975.888268] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 975.889261] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d723623-1115-4d98-be25-0421fddfbce8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.910850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ba6467-b0ac-4ff0-a9c0-20d9430bdc85 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.917787] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4ed467-6aef-41f0-96f1-ee46e234c87d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.937469] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6db0eb2-bdf1-4827-8672-3111d0456921 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.955669] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] The volume has not been displaced from its original location: [datastore2] volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700/volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 975.960814] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.961207] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d116c48b-923f-4530-b22b-a3d4e2de0240 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.982014] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156589, 'name': PowerOffVM_Task, 'duration_secs': 0.446244} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.983306] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 975.983500] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.983836] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 975.983836] env[62066]: value = "task-1156591" [ 975.983836] env[62066]: _type = "Task" [ 975.983836] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.984045] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ad5537b-9c07-4829-95a7-1dacb56788a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.994267] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156591, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.070223] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070597} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.070567] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.071424] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050f269c-627f-4194-9083-e6c9760334d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.091395] env[62066]: DEBUG nova.scheduler.client.report [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.105868] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 045c2efa-e2fe-4d51-b267-e2b380fcf1d2/045c2efa-e2fe-4d51-b267-e2b380fcf1d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.107115] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df3c521c-2053-4c16-a149-e5fa777b7e5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.121429] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.121676] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.122226] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleting the datastore file [datastore1] 10a2f245-93fd-4f6a-8c2f-897fd82269d3 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.122786] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-096bedff-4bc5-475c-b713-730c038b6d9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.134485] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 976.134485] env[62066]: value = "task-1156594" [ 976.134485] env[62066]: _type = "Task" [ 976.134485] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.134892] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for the task: (returnval){ [ 976.134892] env[62066]: value = "task-1156593" [ 976.134892] env[62066]: _type = "Task" [ 976.134892] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.147239] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156593, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.150692] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.335571] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.361973] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.362413] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.362513] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.362690] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.362908] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.363122] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.363390] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.363659] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.363891] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.364131] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.364358] env[62066]: DEBUG nova.virt.hardware [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.365328] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54a52cc-e210-4806-b1b4-9dfbe6640df1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.374451] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5968b89-1b7a-4b41-a2ed-41ed2a46b8ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.496588] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156591, 'name': ReconfigVM_Task, 'duration_secs': 0.316625} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.496825] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.501988] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38f4dfc7-272b-4a65-aa0c-f32979b63f54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.518520] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 976.518520] env[62066]: value = "task-1156595" [ 976.518520] env[62066]: _type = "Task" [ 976.518520] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.529629] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156595, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.607042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.607042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.608406] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.296s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.609029] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 976.612684] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.344s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.613119] env[62066]: DEBUG nova.objects.instance [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 976.652262] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156594, 'name': ReconfigVM_Task, 'duration_secs': 0.293357} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.654960] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 045c2efa-e2fe-4d51-b267-e2b380fcf1d2/045c2efa-e2fe-4d51-b267-e2b380fcf1d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.655878] env[62066]: DEBUG oslo_vmware.api [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Task: {'id': task-1156593, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142086} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.656217] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-650c8543-272c-424b-bc77-0bf3a6f40d7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.658209] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.658428] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.658634] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.658819] env[62066]: INFO nova.compute.manager [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Took 1.22 seconds to destroy the instance on the hypervisor. [ 976.659103] env[62066]: DEBUG oslo.service.loopingcall [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.659402] env[62066]: DEBUG nova.compute.manager [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.659511] env[62066]: DEBUG nova.network.neutron [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.662319] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Successfully updated port: bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.666895] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 976.666895] env[62066]: value = "task-1156596" [ 976.666895] env[62066]: _type = "Task" [ 976.666895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.677160] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156596, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.744844] env[62066]: DEBUG nova.compute.manager [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Received event network-vif-plugged-bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.745894] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Acquiring lock "845748f7-d87b-4141-9932-235acd3d764c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.746312] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Lock "845748f7-d87b-4141-9932-235acd3d764c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.746444] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Lock "845748f7-d87b-4141-9932-235acd3d764c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.746661] env[62066]: DEBUG nova.compute.manager [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] No waiting events found dispatching network-vif-plugged-bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.746977] env[62066]: WARNING nova.compute.manager [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Received unexpected event network-vif-plugged-bc450628-41ad-497c-a20a-4afdc1d4903a for instance with vm_state building and task_state spawning. [ 976.747085] env[62066]: DEBUG nova.compute.manager [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Received event network-changed-bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.747298] env[62066]: DEBUG nova.compute.manager [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Refreshing instance network info cache due to event network-changed-bc450628-41ad-497c-a20a-4afdc1d4903a. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.747546] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Acquiring lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.747770] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Acquired lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.747937] env[62066]: DEBUG nova.network.neutron [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Refreshing network info cache for port bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 977.028747] env[62066]: DEBUG oslo_vmware.api [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156595, 'name': ReconfigVM_Task, 'duration_secs': 0.159642} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.029064] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251748', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'name': 'volume-0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '00f44ecb-768a-4db3-b229-27bb6f27ede1', 'attached_at': '', 'detached_at': '', 'volume_id': '0b5166a2-4e24-4bee-8c6e-0591e1e80700', 'serial': '0b5166a2-4e24-4bee-8c6e-0591e1e80700'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 977.113251] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.116987] env[62066]: DEBUG nova.compute.utils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 977.120958] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.121150] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.161218] env[62066]: DEBUG nova.policy [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aefa9f5bf22f49db846fa171740a687f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1160432c71b042efa6c0e45cf58b37cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.164490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.176818] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156596, 'name': Rename_Task, 'duration_secs': 0.138687} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.177100] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.177347] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31f4420b-a8f0-404b-9fec-4dd7bf69c0db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.186300] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 977.186300] env[62066]: value = "task-1156597" [ 977.186300] env[62066]: _type = "Task" [ 977.186300] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.196058] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.294808] env[62066]: DEBUG nova.network.neutron [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 977.415751] env[62066]: DEBUG nova.network.neutron [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.423027] env[62066]: DEBUG nova.network.neutron [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.482207] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Successfully created port: cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.593025] env[62066]: DEBUG nova.objects.instance [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'flavor' on Instance uuid 00f44ecb-768a-4db3-b229-27bb6f27ede1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.624070] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.628684] env[62066]: DEBUG oslo_concurrency.lockutils [None req-31ae70e8-d50a-40e6-9aab-ac3736a73a84 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.652357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.653052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.654325] env[62066]: INFO nova.compute.claims [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.697606] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.920778] env[62066]: DEBUG oslo_concurrency.lockutils [req-15f1ee29-3257-414e-9cdd-e1edf498af5f req-c8d70358-9c94-4c7a-b686-dfdc61838787 service nova] Releasing lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.921201] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquired lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.921365] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.922647] env[62066]: INFO nova.compute.manager [-] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Took 1.26 seconds to deallocate network for instance. [ 978.197531] env[62066]: DEBUG oslo_vmware.api [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156597, 'name': PowerOnVM_Task, 'duration_secs': 0.704207} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.197832] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.198018] env[62066]: INFO nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Took 8.21 seconds to spawn the instance on the hypervisor. [ 978.198213] env[62066]: DEBUG nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.199045] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05643482-eb3b-4f2b-ba8d-0509456b0927 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.429201] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.452042] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.589613] env[62066]: DEBUG nova.network.neutron [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Updating instance_info_cache with network_info: [{"id": "bc450628-41ad-497c-a20a-4afdc1d4903a", "address": "fa:16:3e:8d:94:d8", "network": {"id": "9fb68368-3f9c-4e8d-9b6d-73a349d1f611", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1121017602-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e6417ca1e084cf5889618dee4e86718", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc450628-41", "ovs_interfaceid": "bc450628-41ad-497c-a20a-4afdc1d4903a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.600818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df6b4d92-79d2-498e-a6c3-4af3ac89d8bf tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.251s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.633585] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 978.658528] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.658787] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.659023] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.659146] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.659299] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.659791] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.659791] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.659791] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.659959] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.661689] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.661689] env[62066]: DEBUG nova.virt.hardware [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.661689] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7250685f-c697-40b3-881e-24ecd670f24a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.672731] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498bbc14-87b0-4a1e-b950-d1c0659759a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.717218] env[62066]: INFO nova.compute.manager [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Took 13.13 seconds to build instance. [ 978.767726] env[62066]: DEBUG nova.compute.manager [req-69e8c238-04b8-479a-832c-4ec31a2bee76 req-786b3722-12c9-4070-a79b-19859e7a108b service nova] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Received event network-vif-deleted-8c87851b-47ba-4dcf-abe5-ab414ef3d4b5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.835812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc67758-043b-4c99-8b63-f82fc697951d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.844794] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2300d6cd-d00e-42dc-b642-222ac3630ffe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.880349] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05df81e-d9ff-467e-88c3-8b29a3eb3449 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.887128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e93dc5-a41f-400b-a4d8-37099471f29b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.901243] env[62066]: DEBUG nova.compute.provider_tree [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.978696] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Successfully updated port: cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.092625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Releasing lock "refresh_cache-845748f7-d87b-4141-9932-235acd3d764c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.092989] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance network_info: |[{"id": "bc450628-41ad-497c-a20a-4afdc1d4903a", "address": "fa:16:3e:8d:94:d8", "network": {"id": "9fb68368-3f9c-4e8d-9b6d-73a349d1f611", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1121017602-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e6417ca1e084cf5889618dee4e86718", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc450628-41", "ovs_interfaceid": "bc450628-41ad-497c-a20a-4afdc1d4903a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.093454] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:94:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc450628-41ad-497c-a20a-4afdc1d4903a', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.101083] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Creating folder: Project (4e6417ca1e084cf5889618dee4e86718). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 979.101366] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e30a7b1-74f4-44cb-9ec0-644ed35d5e3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.112435] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Created folder: Project (4e6417ca1e084cf5889618dee4e86718) in parent group-v251573. [ 979.112636] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Creating folder: Instances. Parent ref: group-v251761. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 979.112888] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddb85c3b-459d-4f9c-a717-f4f013352931 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.122041] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Created folder: Instances in parent group-v251761. [ 979.122300] env[62066]: DEBUG oslo.service.loopingcall [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.122527] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.122767] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d433ec4-e662-4a71-9244-4d9e62d3778e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.142863] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.142863] env[62066]: value = "task-1156601" [ 979.142863] env[62066]: _type = "Task" [ 979.142863] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.150208] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156601, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.219120] env[62066]: DEBUG oslo_concurrency.lockutils [None req-52358279-2527-4c94-97fc-92ecdf212c51 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.640s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.404094] env[62066]: DEBUG nova.scheduler.client.report [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.485023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.485023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.485023] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.599147] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 979.599410] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251760', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'name': 'volume-9efbca39-3380-44d5-a311-1c9255ac3a78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138', 'attached_at': '', 'detached_at': '', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'serial': '9efbca39-3380-44d5-a311-1c9255ac3a78'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 979.600319] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5659bd2f-1b06-4846-9812-c3d5a00328f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.618281] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdad933-2598-40e5-bdbb-21dc6ffd850e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.645940] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] volume-9efbca39-3380-44d5-a311-1c9255ac3a78/volume-9efbca39-3380-44d5-a311-1c9255ac3a78.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.647395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ce56ecc-0aa0-4d5e-a5df-f5ffee36dd79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.667851] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156601, 'name': CreateVM_Task, 'duration_secs': 0.332282} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.670188] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.670188] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 979.670188] env[62066]: value = "task-1156602" [ 979.670188] env[62066]: _type = "Task" [ 979.670188] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.670483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.670651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.670971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.671272] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07418a68-2422-44bd-9707-ec71b5b352ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.679337] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 979.679337] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528120b0-7516-a06c-911c-96f1dcb3fa43" [ 979.679337] env[62066]: _type = "Task" [ 979.679337] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.684478] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156602, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.692959] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528120b0-7516-a06c-911c-96f1dcb3fa43, 'name': SearchDatastore_Task, 'duration_secs': 0.010929} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.694217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.694217] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.694217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.694469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.694621] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.694898] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-726c60ca-589f-4230-8d06-8b0301356982 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.704134] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.704134] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.704441] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-708698dd-687b-41b3-9306-05e943083caf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.709467] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 979.709467] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5237c1a2-6488-bc62-0727-01c237720b3c" [ 979.709467] env[62066]: _type = "Task" [ 979.709467] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.717178] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5237c1a2-6488-bc62-0727-01c237720b3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.768471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.768742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.769020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.769237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.769410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.772027] env[62066]: INFO nova.compute.manager [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Terminating instance [ 979.773750] env[62066]: DEBUG nova.compute.manager [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.773959] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.775166] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c46144-6c97-43c1-bc09-23b934916923 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.784256] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.784506] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78568c47-a257-4cd9-affc-32c4c1a67668 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.791630] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 979.791630] env[62066]: value = "task-1156603" [ 979.791630] env[62066]: _type = "Task" [ 979.791630] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.799882] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.909151] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.909745] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.912564] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.484s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.916056] env[62066]: DEBUG nova.objects.instance [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lazy-loading 'resources' on Instance uuid 10a2f245-93fd-4f6a-8c2f-897fd82269d3 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.025470] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.183281] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156602, 'name': ReconfigVM_Task, 'duration_secs': 0.384538} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.183467] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to attach disk [datastore2] volume-9efbca39-3380-44d5-a311-1c9255ac3a78/volume-9efbca39-3380-44d5-a311-1c9255ac3a78.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.188285] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1745e573-6a40-426b-82b9-a27070b83395 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.200647] env[62066]: DEBUG nova.network.neutron [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Updating instance_info_cache with network_info: [{"id": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "address": "fa:16:3e:6f:5a:9b", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6f245e-66", "ovs_interfaceid": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.208028] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 980.208028] env[62066]: value = "task-1156604" [ 980.208028] env[62066]: _type = "Task" [ 980.208028] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.223294] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156604, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.227120] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5237c1a2-6488-bc62-0727-01c237720b3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008622} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.228283] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c59f8e1-a3d6-4f37-bd64-6eca838ab2c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.234175] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 980.234175] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524c37ce-40ba-bf01-1fa5-25965d029b4a" [ 980.234175] env[62066]: _type = "Task" [ 980.234175] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.246434] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524c37ce-40ba-bf01-1fa5-25965d029b4a, 'name': SearchDatastore_Task, 'duration_secs': 0.009663} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.246805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.247179] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 845748f7-d87b-4141-9932-235acd3d764c/845748f7-d87b-4141-9932-235acd3d764c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.247525] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ba550e0-c891-4c1c-b15f-196ad07f1105 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.256726] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 980.256726] env[62066]: value = "task-1156605" [ 980.256726] env[62066]: _type = "Task" [ 980.256726] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.266332] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.303435] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156603, 'name': PowerOffVM_Task, 'duration_secs': 0.222888} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.303678] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.303897] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.304146] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2d42e99-43b5-4cc1-8267-603a837359ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.363833] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.363833] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.363833] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore2] 00f44ecb-768a-4db3-b229-27bb6f27ede1 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.363982] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4a36c7e-dda5-48ba-bfbc-109a142fb516 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.370139] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 980.370139] env[62066]: value = "task-1156607" [ 980.370139] env[62066]: _type = "Task" [ 980.370139] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.379063] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156607, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.415642] env[62066]: DEBUG nova.compute.utils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 980.417338] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.417545] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.496728] env[62066]: DEBUG nova.policy [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.622562] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974a60e8-b5b6-4f99-a711-13de09dd083d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.636537] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51f7bcc-5d4f-458f-b785-f7b4ef0b7f18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.682468] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d1d2c7-3f3e-4515-acaa-ca7c68e8e70c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.691561] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a89bd55-2b1a-49df-b84f-9f21c3d756d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.709474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.709936] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Instance network_info: |[{"id": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "address": "fa:16:3e:6f:5a:9b", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6f245e-66", "ovs_interfaceid": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 980.710694] env[62066]: DEBUG nova.compute.provider_tree [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.713054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:5a:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89470f7f-1c8b-4c83-92b5-6f73a77c520f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc6f245e-66f3-4fdb-8458-a44d45b6e8af', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.721906] env[62066]: DEBUG oslo.service.loopingcall [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.726882] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 980.727301] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d376395-8164-49f7-a372-dc99c72fb696 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.754024] env[62066]: DEBUG oslo_vmware.api [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156604, 'name': ReconfigVM_Task, 'duration_secs': 0.143854} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.754024] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251760', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'name': 'volume-9efbca39-3380-44d5-a311-1c9255ac3a78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138', 'attached_at': '', 'detached_at': '', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'serial': '9efbca39-3380-44d5-a311-1c9255ac3a78'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 980.755800] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.755800] env[62066]: value = "task-1156608" [ 980.755800] env[62066]: _type = "Task" [ 980.755800] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.771895] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156608, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.778149] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156605, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.812878] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Received event network-vif-plugged-cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.813127] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquiring lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.813348] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.813523] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.813737] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] No waiting events found dispatching network-vif-plugged-cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.813947] env[62066]: WARNING nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Received unexpected event network-vif-plugged-cc6f245e-66f3-4fdb-8458-a44d45b6e8af for instance with vm_state building and task_state spawning. [ 980.814106] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Received event network-changed-cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.814296] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Refreshing instance network info cache due to event network-changed-cc6f245e-66f3-4fdb-8458-a44d45b6e8af. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.814750] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquiring lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.815855] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquired lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.815855] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Refreshing network info cache for port cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.880683] env[62066]: DEBUG oslo_vmware.api [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156607, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.448777} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.880918] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.881127] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.881317] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.881514] env[62066]: INFO nova.compute.manager [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 980.881785] env[62066]: DEBUG oslo.service.loopingcall [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.881989] env[62066]: DEBUG nova.compute.manager [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.882099] env[62066]: DEBUG nova.network.neutron [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.924477] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.213845] env[62066]: DEBUG nova.scheduler.client.report [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.272531] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535558} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.275859] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 845748f7-d87b-4141-9932-235acd3d764c/845748f7-d87b-4141-9932-235acd3d764c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.276145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.277026] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156608, 'name': CreateVM_Task, 'duration_secs': 0.361347} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.277026] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca3503e5-5fde-4393-b218-61aabf5acf65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.279452] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 981.280028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.280028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.280358] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.282161] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce6cbc89-ab2c-4c39-8ddd-80e627f39c3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.288119] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 981.288119] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b20afc-bc1a-64ca-c058-ee59417f554f" [ 981.288119] env[62066]: _type = "Task" [ 981.288119] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.288119] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 981.288119] env[62066]: value = "task-1156609" [ 981.288119] env[62066]: _type = "Task" [ 981.288119] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.303117] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b20afc-bc1a-64ca-c058-ee59417f554f, 'name': SearchDatastore_Task, 'duration_secs': 0.010761} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.306275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.307118] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.307118] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.307118] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.307118] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.307451] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.308317] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Successfully created port: e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.310114] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2e27e5f-5dff-4740-a372-876a22ca4624 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.321859] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.322443] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 981.323441] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b233f4b1-d7bf-4bc6-b0f5-2de4845cf347 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.329230] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 981.329230] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e947c6-99d6-b9b1-8f7b-d1c39a790618" [ 981.329230] env[62066]: _type = "Task" [ 981.329230] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.337935] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e947c6-99d6-b9b1-8f7b-d1c39a790618, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.710907] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Updated VIF entry in instance network info cache for port cc6f245e-66f3-4fdb-8458-a44d45b6e8af. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 981.711616] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Updating instance_info_cache with network_info: [{"id": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "address": "fa:16:3e:6f:5a:9b", "network": {"id": "54f564a8-0548-4f4e-8c1c-e5814506a17d", "bridge": "br-int", "label": "tempest-ServersTestJSON-796542756-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1160432c71b042efa6c0e45cf58b37cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89470f7f-1c8b-4c83-92b5-6f73a77c520f", "external-id": "nsx-vlan-transportzone-929", "segmentation_id": 929, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6f245e-66", "ovs_interfaceid": "cc6f245e-66f3-4fdb-8458-a44d45b6e8af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.718173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.806s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.750258] env[62066]: INFO nova.scheduler.client.report [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Deleted allocations for instance 10a2f245-93fd-4f6a-8c2f-897fd82269d3 [ 981.807655] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.292101} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.807855] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.809745] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475e7909-9a11-45d3-a30a-4e129a92798a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.827361] env[62066]: DEBUG nova.objects.instance [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'flavor' on Instance uuid ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.843328] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 845748f7-d87b-4141-9932-235acd3d764c/845748f7-d87b-4141-9932-235acd3d764c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.843328] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b3f29b7-6717-4e5d-830e-220948b21ee1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.868733] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e947c6-99d6-b9b1-8f7b-d1c39a790618, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.874294] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 981.874294] env[62066]: value = "task-1156610" [ 981.874294] env[62066]: _type = "Task" [ 981.874294] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.874294] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-080403a6-10f9-42b7-89f4-5956a5d10177 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.879260] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 981.879260] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528fc433-aef5-56a3-96e0-20992cad2f6a" [ 981.879260] env[62066]: _type = "Task" [ 981.879260] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.882674] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156610, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.892702] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528fc433-aef5-56a3-96e0-20992cad2f6a, 'name': SearchDatastore_Task, 'duration_secs': 0.009498} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.893029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.893353] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 7da117c1-0686-4b7e-9c33-5935a103dc3e/7da117c1-0686-4b7e-9c33-5935a103dc3e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.894025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c0b2dc4-4dc9-48c7-b756-1b5143214955 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.900421] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 981.900421] env[62066]: value = "task-1156611" [ 981.900421] env[62066]: _type = "Task" [ 981.900421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.909165] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.937063] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.972690] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.972953] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.973126] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.973311] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.973459] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.973610] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.973866] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.974202] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.974446] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.974624] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.974843] env[62066]: DEBUG nova.virt.hardware [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.975995] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8eae4f-2533-4500-8292-6a8482362220 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.984528] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6158c04d-cb41-486d-84ec-bb67e9121c5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.217683] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Releasing lock "refresh_cache-7da117c1-0686-4b7e-9c33-5935a103dc3e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.217683] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.217683] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.217683] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.217683] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.217683] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 982.265239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0cef34db-5c6f-409f-91f0-6565ae48508b tempest-ServerDiskConfigTestJSON-1076736961 tempest-ServerDiskConfigTestJSON-1076736961-project-member] Lock "10a2f245-93fd-4f6a-8c2f-897fd82269d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.836s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.347044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6e40c372-44ff-4d27-80f9-60db179b7312 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.347s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.383227] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156610, 'name': ReconfigVM_Task, 'duration_secs': 0.281202} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.383519] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 845748f7-d87b-4141-9932-235acd3d764c/845748f7-d87b-4141-9932-235acd3d764c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.384267] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b2dddc9-2528-48de-a770-76d1c3b62e71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.392156] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 982.392156] env[62066]: value = "task-1156612" [ 982.392156] env[62066]: _type = "Task" [ 982.392156] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.400300] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156612, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.408867] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156611, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491118} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.409112] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 7da117c1-0686-4b7e-9c33-5935a103dc3e/7da117c1-0686-4b7e-9c33-5935a103dc3e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.409330] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.409563] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d48fb86-11a4-4d48-bfec-dab104d6f39c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.415810] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 982.415810] env[62066]: value = "task-1156613" [ 982.415810] env[62066]: _type = "Task" [ 982.415810] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.425412] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.520800] env[62066]: DEBUG nova.network.neutron [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.857690] env[62066]: DEBUG nova.compute.manager [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.857690] env[62066]: DEBUG nova.compute.manager [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing instance network info cache due to event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.857690] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.857690] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.857971] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 982.905976] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156612, 'name': Rename_Task, 'duration_secs': 0.163404} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.906172] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.906453] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8784b69-a9d0-4785-989e-777dd668a270 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.912914] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 982.912914] env[62066]: value = "task-1156614" [ 982.912914] env[62066]: _type = "Task" [ 982.912914] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.924473] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063778} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.927357] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.928045] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.928747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d74713-97d1-455f-8e31-78b5043b8c7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.952259] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 7da117c1-0686-4b7e-9c33-5935a103dc3e/7da117c1-0686-4b7e-9c33-5935a103dc3e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.953281] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.953541] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.955950] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1939b2be-9d04-4045-8d60-56395e7dfa29 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.975041] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 982.975041] env[62066]: value = "task-1156615" [ 982.975041] env[62066]: _type = "Task" [ 982.975041] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.983565] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156615, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.022562] env[62066]: INFO nova.compute.manager [-] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Took 2.14 seconds to deallocate network for instance. [ 983.219074] env[62066]: DEBUG nova.compute.manager [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 983.283452] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Successfully updated port: e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.421819] env[62066]: DEBUG oslo_vmware.api [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156614, 'name': PowerOnVM_Task, 'duration_secs': 0.473387} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.422136] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.422342] env[62066]: INFO nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Took 7.09 seconds to spawn the instance on the hypervisor. [ 983.422526] env[62066]: DEBUG nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.423852] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a700f961-f70f-49c2-b343-15d67c331187 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.470876] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.471183] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.471357] env[62066]: DEBUG nova.compute.manager [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing instance network info cache due to event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.471552] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.486190] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156615, 'name': ReconfigVM_Task, 'duration_secs': 0.294432} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.486484] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 7da117c1-0686-4b7e-9c33-5935a103dc3e/7da117c1-0686-4b7e-9c33-5935a103dc3e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.487115] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2638bb83-61f4-41fe-b58b-fd50b1d11714 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.493635] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 983.493635] env[62066]: value = "task-1156616" [ 983.493635] env[62066]: _type = "Task" [ 983.493635] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.501702] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156616, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.530743] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.530965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.531211] env[62066]: DEBUG nova.objects.instance [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'resources' on Instance uuid 00f44ecb-768a-4db3-b229-27bb6f27ede1 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.578343] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updated VIF entry in instance network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.578906] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.737800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.791718] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.792016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.792212] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.940039] env[62066]: INFO nova.compute.manager [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Took 13.42 seconds to build instance. [ 984.004152] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156616, 'name': Rename_Task, 'duration_secs': 0.136406} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.004461] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.004704] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b26edd79-a869-48b4-a5b1-5794b1b45412 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.012358] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 984.012358] env[62066]: value = "task-1156617" [ 984.012358] env[62066]: _type = "Task" [ 984.012358] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.021710] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.082382] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.082661] env[62066]: DEBUG nova.compute.manager [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.082838] env[62066]: DEBUG nova.compute.manager [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.083044] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.083198] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.083363] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.084605] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.084789] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.214959] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c1546f-0bb1-43f9-8aee-3713fa35c8b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.223213] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed2f78d-f660-4e57-ab50-21ef11c4a95d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.256175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0dc575-754e-449c-83ed-b4230df8c176 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.264597] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb71d0b-dd27-4c6c-88c6-3067a1850b1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.279974] env[62066]: DEBUG nova.compute.provider_tree [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.329208] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.442190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9b9f9cb4-edea-4f4a-b5f8-1c316ec7aacf tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.939s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.479655] env[62066]: DEBUG nova.network.neutron [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Updating instance_info_cache with network_info: [{"id": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "address": "fa:16:3e:19:d2:db", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c0d760-32", "ovs_interfaceid": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.523034] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156617, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.783293] env[62066]: DEBUG nova.scheduler.client.report [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.917188] env[62066]: DEBUG nova.compute.manager [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Received event network-vif-plugged-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.917360] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Acquiring lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.917551] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.917930] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.918134] env[62066]: DEBUG nova.compute.manager [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] No waiting events found dispatching network-vif-plugged-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.918310] env[62066]: WARNING nova.compute.manager [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Received unexpected event network-vif-plugged-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac for instance with vm_state building and task_state spawning. [ 984.918474] env[62066]: DEBUG nova.compute.manager [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Received event network-changed-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.918864] env[62066]: DEBUG nova.compute.manager [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Refreshing instance network info cache due to event network-changed-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 984.919066] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Acquiring lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.940844] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updated VIF entry in instance network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 984.941413] env[62066]: DEBUG nova.network.neutron [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.963422] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 984.963824] env[62066]: DEBUG nova.network.neutron [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.983169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.983169] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Instance network_info: |[{"id": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "address": "fa:16:3e:19:d2:db", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c0d760-32", "ovs_interfaceid": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.983358] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Acquired lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.983579] env[62066]: DEBUG nova.network.neutron [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Refreshing network info cache for port e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.988153] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:d2:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1c0d760-32cf-44fb-b45e-0aa1b25c47ac', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.000612] env[62066]: DEBUG oslo.service.loopingcall [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.005150] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.005798] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83368b4b-d84a-4ff4-a311-1090864342b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.043018] env[62066]: DEBUG oslo_vmware.api [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156617, 'name': PowerOnVM_Task, 'duration_secs': 0.806446} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.044545] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.044879] env[62066]: INFO nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Took 6.41 seconds to spawn the instance on the hypervisor. [ 985.045166] env[62066]: DEBUG nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.045882] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.045882] env[62066]: value = "task-1156618" [ 985.045882] env[62066]: _type = "Task" [ 985.045882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.046894] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0212c7a-1f0a-4e3a-b158-d9a1fd1b3900 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.065549] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156618, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.214827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "845748f7-d87b-4141-9932-235acd3d764c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.215081] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.215318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "845748f7-d87b-4141-9932-235acd3d764c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.215518] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.215731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.218140] env[62066]: INFO nova.compute.manager [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Terminating instance [ 985.219654] env[62066]: DEBUG nova.compute.manager [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.219849] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.220681] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c87a391-7f1d-486c-b5b1-ff8ebf107efe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.228791] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.229055] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f929ef15-7f8f-4089-ab5a-a12112688078 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.235467] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 985.235467] env[62066]: value = "task-1156619" [ 985.235467] env[62066]: _type = "Task" [ 985.235467] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.244193] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.289633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.292746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.555s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.309302] env[62066]: DEBUG nova.network.neutron [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Updated VIF entry in instance network info cache for port e1c0d760-32cf-44fb-b45e-0aa1b25c47ac. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.310408] env[62066]: DEBUG nova.network.neutron [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Updating instance_info_cache with network_info: [{"id": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "address": "fa:16:3e:19:d2:db", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1c0d760-32", "ovs_interfaceid": "e1c0d760-32cf-44fb-b45e-0aa1b25c47ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.320394] env[62066]: INFO nova.scheduler.client.report [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocations for instance 00f44ecb-768a-4db3-b229-27bb6f27ede1 [ 985.444050] env[62066]: DEBUG oslo_concurrency.lockutils [req-56c5569c-a672-471e-b7e5-aba3ad3b9fce req-085130ad-2213-4322-b7d6-047148643cee service nova] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.466560] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.466847] env[62066]: DEBUG nova.compute.manager [req-cc01ae6b-6863-4423-8941-2b3853a7ac3f req-3726682e-0fd2-46d7-8454-c0c778f9cc80 service nova] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Received event network-vif-deleted-f0fae426-1b77-435b-98a6-72d487ff589f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.560728] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156618, 'name': CreateVM_Task, 'duration_secs': 0.379193} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.560928] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.562012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.562012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.562275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.562552] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17ed556f-017f-4cda-8d7d-518b5017c91e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.569612] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 985.569612] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527cdb71-d1bb-058f-e1c5-67306c7d61d1" [ 985.569612] env[62066]: _type = "Task" [ 985.569612] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.590374] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527cdb71-d1bb-058f-e1c5-67306c7d61d1, 'name': SearchDatastore_Task, 'duration_secs': 0.009011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.591588] env[62066]: INFO nova.compute.manager [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Took 12.65 seconds to build instance. [ 985.591842] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.592101] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.592343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.592494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.592674] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.593358] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e70025f0-73fa-42ef-89e6-14b61fe7af8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.601788] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.602027] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.602706] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-406a3316-7848-4688-9333-cf37dd5a0ef0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.607951] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 985.607951] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52707c75-feef-d3bb-7f00-6e65d93bbf2d" [ 985.607951] env[62066]: _type = "Task" [ 985.607951] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.619364] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52707c75-feef-d3bb-7f00-6e65d93bbf2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.745630] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156619, 'name': PowerOffVM_Task, 'duration_secs': 0.18559} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.745910] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.746123] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.746388] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55db1a50-6bdb-4468-8d95-74d5cee9df20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.798897] env[62066]: INFO nova.compute.claims [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.814596] env[62066]: DEBUG oslo_concurrency.lockutils [req-6dc352c4-6f8a-4613-a89b-819480f278a4 req-1ca3476e-d399-41b3-9843-e0d1a894be84 service nova] Releasing lock "refresh_cache-7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.816481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.816481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.816481] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Deleting the datastore file [datastore2] 845748f7-d87b-4141-9932-235acd3d764c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.816481] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a94d23d9-ae08-4fc9-a5ef-8fab375c5695 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.822251] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for the task: (returnval){ [ 985.822251] env[62066]: value = "task-1156621" [ 985.822251] env[62066]: _type = "Task" [ 985.822251] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.829935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c726222d-f663-47ae-99fc-386d17cf07f8 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "00f44ecb-768a-4db3-b229-27bb6f27ede1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.061s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.835506] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.096553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b7f76b0f-d750-4139-8b6b-893c095668c3 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.160s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.120102] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52707c75-feef-d3bb-7f00-6e65d93bbf2d, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.120225] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b1203e4-82d1-4c3d-b7f0-c8174877caab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.126733] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 986.126733] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525f1d2c-1ee8-ce02-26cd-4c4841866bf2" [ 986.126733] env[62066]: _type = "Task" [ 986.126733] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.134968] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525f1d2c-1ee8-ce02-26cd-4c4841866bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.305983] env[62066]: INFO nova.compute.resource_tracker [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating resource usage from migration 02ce68b7-a528-4678-aa25-7fe4e6bfb631 [ 986.337038] env[62066]: DEBUG oslo_vmware.api [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Task: {'id': task-1156621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201926} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.337308] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.337642] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.337901] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.338329] env[62066]: INFO nova.compute.manager [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 986.338429] env[62066]: DEBUG oslo.service.loopingcall [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.340988] env[62066]: DEBUG nova.compute.manager [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.341123] env[62066]: DEBUG nova.network.neutron [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.463620] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729c0ce5-d144-4913-af0a-4b8791f27924 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.472294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3748feb-e007-420f-b7dd-0463f349ca5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.505862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced806c7-6673-467e-a44a-87b32f548908 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.513444] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da580f2-3f89-4b04-9333-b14c88d1e165 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.526826] env[62066]: DEBUG nova.compute.provider_tree [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.637841] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525f1d2c-1ee8-ce02-26cd-4c4841866bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.638078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.638359] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc/7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.638635] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e5e7823-962c-4f9b-8a88-af154e69407c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.645723] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 986.645723] env[62066]: value = "task-1156622" [ 986.645723] env[62066]: _type = "Task" [ 986.645723] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.653736] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156622, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.931049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.931311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.931480] env[62066]: DEBUG nova.compute.manager [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.932379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c74d6f7-c846-4b50-ab54-207351fcb155 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.939656] env[62066]: DEBUG nova.compute.manager [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 986.940334] env[62066]: DEBUG nova.objects.instance [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'flavor' on Instance uuid 7da117c1-0686-4b7e-9c33-5935a103dc3e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.944338] env[62066]: DEBUG nova.compute.manager [req-d41b9f58-481b-433a-aebb-6daef6e0cb79 req-0381fbd4-f87a-40f5-b687-05a145eaeabd service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Received event network-vif-deleted-bc450628-41ad-497c-a20a-4afdc1d4903a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.944539] env[62066]: INFO nova.compute.manager [req-d41b9f58-481b-433a-aebb-6daef6e0cb79 req-0381fbd4-f87a-40f5-b687-05a145eaeabd service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Neutron deleted interface bc450628-41ad-497c-a20a-4afdc1d4903a; detaching it from the instance and deleting it from the info cache [ 986.944715] env[62066]: DEBUG nova.network.neutron [req-d41b9f58-481b-433a-aebb-6daef6e0cb79 req-0381fbd4-f87a-40f5-b687-05a145eaeabd service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.030472] env[62066]: DEBUG nova.scheduler.client.report [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.091742] env[62066]: DEBUG nova.network.neutron [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.157059] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156622, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.448862] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-893e2ff2-3a04-4c8f-a2a0-1649998b891b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.451292] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.451554] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59d65c47-806f-47de-b0f8-104ae4ede2fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.461576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f5c061-b438-484b-b0e1-03613db860cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.474635] env[62066]: DEBUG oslo_vmware.api [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 987.474635] env[62066]: value = "task-1156623" [ 987.474635] env[62066]: _type = "Task" [ 987.474635] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.485369] env[62066]: DEBUG oslo_vmware.api [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156623, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.494977] env[62066]: DEBUG nova.compute.manager [req-d41b9f58-481b-433a-aebb-6daef6e0cb79 req-0381fbd4-f87a-40f5-b687-05a145eaeabd service nova] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Detach interface failed, port_id=bc450628-41ad-497c-a20a-4afdc1d4903a, reason: Instance 845748f7-d87b-4141-9932-235acd3d764c could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 987.534469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.242s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.534693] env[62066]: INFO nova.compute.manager [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Migrating [ 987.593544] env[62066]: INFO nova.compute.manager [-] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Took 1.25 seconds to deallocate network for instance. [ 987.657204] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156622, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.925842} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.657367] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc/7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.657636] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.657935] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84d03afc-de10-4252-b791-8c7fa1ee6a84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.666400] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 987.666400] env[62066]: value = "task-1156624" [ 987.666400] env[62066]: _type = "Task" [ 987.666400] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.677253] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.984393] env[62066]: DEBUG oslo_vmware.api [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156623, 'name': PowerOffVM_Task, 'duration_secs': 0.242188} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.984667] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.984865] env[62066]: DEBUG nova.compute.manager [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.985627] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae62d9e-120c-4be9-940f-2184fa0c82dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.049931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.050137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.050321] env[62066]: DEBUG nova.network.neutron [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.100014] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.100329] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.100571] env[62066]: DEBUG nova.objects.instance [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lazy-loading 'resources' on Instance uuid 845748f7-d87b-4141-9932-235acd3d764c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.179940] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.180270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.181907] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076784} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.181982] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.182849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f24425-c8a8-4fe8-8053-b69deb511de1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.205136] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc/7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.205855] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbbe48e3-77e3-4aa7-839f-61958492bd79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.226619] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 988.226619] env[62066]: value = "task-1156625" [ 988.226619] env[62066]: _type = "Task" [ 988.226619] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.234983] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.497763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8a933fb2-6a41-4b2d-8837-60bae9f03243 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.566s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.682420] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 988.740214] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156625, 'name': ReconfigVM_Task, 'duration_secs': 0.296296} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.740214] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc/7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.741054] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3beafba7-d9a9-43ff-a2b6-c2ebb65f0dfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.749142] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 988.749142] env[62066]: value = "task-1156626" [ 988.749142] env[62066]: _type = "Task" [ 988.749142] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.759335] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156626, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.772688] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb04982-b1ad-451c-986d-b0ba43e61b13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.779136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc93019-a8bd-423b-9736-11a1007d8116 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.811088] env[62066]: DEBUG nova.network.neutron [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.814026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d205837-07a3-460a-a5cb-9b191e18142a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.822170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6dee03-eac1-41d0-9d01-579df6834b19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.837487] env[62066]: DEBUG nova.compute.provider_tree [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.202939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.259347] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156626, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.317146] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.340273] env[62066]: DEBUG nova.scheduler.client.report [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.668639] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.668989] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.669166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.669440] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.669662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.672016] env[62066]: INFO nova.compute.manager [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Terminating instance [ 989.674068] env[62066]: DEBUG nova.compute.manager [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.674282] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.675167] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d144a86-6dfb-45a3-a2f3-f10d76002cb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.682970] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.683220] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8be072c7-08a8-4c32-9ee7-3c3a208bd814 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.755064] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.755305] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.755488] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] 7da117c1-0686-4b7e-9c33-5935a103dc3e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.756184] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97cbb92f-2430-4439-82e3-c09ead7d5fb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.760237] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156626, 'name': Rename_Task, 'duration_secs': 0.850669} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.760828] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.761068] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea834441-9423-4740-bbb2-7f427ed4591a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.764128] env[62066]: DEBUG oslo_vmware.api [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 989.764128] env[62066]: value = "task-1156628" [ 989.764128] env[62066]: _type = "Task" [ 989.764128] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.767942] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 989.767942] env[62066]: value = "task-1156629" [ 989.767942] env[62066]: _type = "Task" [ 989.767942] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.773410] env[62066]: DEBUG oslo_vmware.api [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.777668] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.845078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.847903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.645s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.849279] env[62066]: INFO nova.compute.claims [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 989.863799] env[62066]: INFO nova.scheduler.client.report [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Deleted allocations for instance 845748f7-d87b-4141-9932-235acd3d764c [ 990.277351] env[62066]: DEBUG oslo_vmware.api [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131466} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.280202] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.280402] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.280595] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.280772] env[62066]: INFO nova.compute.manager [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 990.281104] env[62066]: DEBUG oslo.service.loopingcall [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.281224] env[62066]: DEBUG oslo_vmware.api [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156629, 'name': PowerOnVM_Task, 'duration_secs': 0.415749} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.281420] env[62066]: DEBUG nova.compute.manager [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.281514] env[62066]: DEBUG nova.network.neutron [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 990.283101] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.283317] env[62066]: INFO nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Took 8.35 seconds to spawn the instance on the hypervisor. [ 990.283502] env[62066]: DEBUG nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.284253] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81541b20-d91a-4bd5-b43a-84c1ffd0834e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.371843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cea3123b-8cfc-4785-a261-90f8e0480b22 tempest-ServersNegativeTestMultiTenantJSON-1870779430 tempest-ServersNegativeTestMultiTenantJSON-1870779430-project-member] Lock "845748f7-d87b-4141-9932-235acd3d764c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.157s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.559858] env[62066]: DEBUG nova.compute.manager [req-8565361b-caed-445e-8cd0-ba08591f5ead req-d50d2b31-03ad-4130-838b-496256cd3912 service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Received event network-vif-deleted-cc6f245e-66f3-4fdb-8458-a44d45b6e8af {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.560117] env[62066]: INFO nova.compute.manager [req-8565361b-caed-445e-8cd0-ba08591f5ead req-d50d2b31-03ad-4130-838b-496256cd3912 service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Neutron deleted interface cc6f245e-66f3-4fdb-8458-a44d45b6e8af; detaching it from the instance and deleting it from the info cache [ 990.560381] env[62066]: DEBUG nova.network.neutron [req-8565361b-caed-445e-8cd0-ba08591f5ead req-d50d2b31-03ad-4130-838b-496256cd3912 service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.800325] env[62066]: INFO nova.compute.manager [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Took 13.17 seconds to build instance. [ 990.834019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6792814c-c61e-4d46-8da3-e6ae06ac0c4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.854134] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.018786] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8000862-2334-4f58-8209-31c28e6bad9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.027974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb72fe1e-4258-426f-bab5-126e9ee30e90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.033436] env[62066]: DEBUG nova.network.neutron [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.061806] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365d4fd1-0034-4689-9582-9986ecda29a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.066701] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2626973-4a98-4907-9f77-d5cb6eb1b7a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.074537] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991e347d-3482-4e63-9222-aef0a29efc30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.081132] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01ab623-8b2f-4586-bb23-a9707657d46f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.106227] env[62066]: DEBUG nova.compute.provider_tree [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.120910] env[62066]: DEBUG nova.compute.manager [req-8565361b-caed-445e-8cd0-ba08591f5ead req-d50d2b31-03ad-4130-838b-496256cd3912 service nova] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Detach interface failed, port_id=cc6f245e-66f3-4fdb-8458-a44d45b6e8af, reason: Instance 7da117c1-0686-4b7e-9c33-5935a103dc3e could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.122115] env[62066]: DEBUG nova.scheduler.client.report [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.302587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-62c63357-3850-47cb-8293-a6f0e5e9a48b tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.696s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.360175] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.360492] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71b65adc-32c7-47ff-b78e-56ce58e23e43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.367724] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 991.367724] env[62066]: value = "task-1156630" [ 991.367724] env[62066]: _type = "Task" [ 991.367724] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.377653] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.523559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.523559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.523559] env[62066]: DEBUG nova.compute.manager [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.524890] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18dfd107-df49-45b9-aa1f-6107d2c1c85b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.531397] env[62066]: DEBUG nova.compute.manager [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 991.532024] env[62066]: DEBUG nova.objects.instance [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'flavor' on Instance uuid 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.536216] env[62066]: INFO nova.compute.manager [-] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Took 1.25 seconds to deallocate network for instance. [ 991.626644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.779s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.627226] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.879296] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156630, 'name': PowerOffVM_Task, 'duration_secs': 0.231616} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.879613] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.879799] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.036871] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.037162] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2b36ce9-5b6c-4e68-bf45-9a0cdaceedb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.041260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.041492] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.041697] env[62066]: DEBUG nova.objects.instance [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'resources' on Instance uuid 7da117c1-0686-4b7e-9c33-5935a103dc3e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.044895] env[62066]: DEBUG oslo_vmware.api [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 992.044895] env[62066]: value = "task-1156631" [ 992.044895] env[62066]: _type = "Task" [ 992.044895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.053830] env[62066]: DEBUG oslo_vmware.api [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.131862] env[62066]: DEBUG nova.compute.utils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 992.133387] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.137134] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 992.172198] env[62066]: DEBUG nova.policy [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91585bdab75c42178c835de912fd3fef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08737e043ede43b58298d1c1a834fa84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.386472] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.386739] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.386904] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.387108] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.387296] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.387495] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.387722] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.387887] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.388606] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.388828] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.389027] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.394854] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20c93ab2-32aa-4c91-b41f-b78e09c0a4be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.414108] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 992.414108] env[62066]: value = "task-1156632" [ 992.414108] env[62066]: _type = "Task" [ 992.414108] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.420827] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156632, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.438931] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Successfully created port: d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.566145] env[62066]: DEBUG oslo_vmware.api [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156631, 'name': PowerOffVM_Task, 'duration_secs': 0.170517} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.566548] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.566763] env[62066]: DEBUG nova.compute.manager [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.567887] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4b754c-4f41-488d-b626-7f4871a24908 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.637129] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.768877] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f9d109-566e-4501-a6f8-9d288e0eaa9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.776139] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbfdde9-e735-483a-934e-d07229ef979f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.816489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c1e8f2-6fcc-4b68-b99d-2d85507fda7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.826347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b995c6-e2a6-464e-8322-a75393bc005c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.842295] env[62066]: DEBUG nova.compute.provider_tree [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.923223] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156632, 'name': ReconfigVM_Task, 'duration_secs': 0.239222} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.923630] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.088422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-75f2718a-e595-4f77-ada8-9c7b781e1e6e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.348461] env[62066]: DEBUG nova.scheduler.client.report [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.353270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "16390d89-f2d1-4d6c-be62-b85e61906865" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.353512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.353726] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.354357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.354544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.356545] env[62066]: INFO nova.compute.manager [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Terminating instance [ 993.358396] env[62066]: DEBUG nova.compute.manager [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.358599] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.359499] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67290ba1-08e2-4063-a17d-eaf384424c95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.368587] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.368587] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ebb2283-f185-4deb-976e-3e7419193d0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.378998] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 993.378998] env[62066]: value = "task-1156633" [ 993.378998] env[62066]: _type = "Task" [ 993.378998] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.390385] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.432610] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.432750] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.433561] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.433561] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.433561] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.433561] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.433939] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.433939] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.434042] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.435029] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.435029] env[62066]: DEBUG nova.virt.hardware [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.439838] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 993.440544] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-352fee2d-ed7b-404e-adf7-d4abcc62fb0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.459980] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 993.459980] env[62066]: value = "task-1156634" [ 993.459980] env[62066]: _type = "Task" [ 993.459980] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.469209] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.646771] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.673171] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.673482] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.673594] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.673766] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.673930] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.674110] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.674325] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.674483] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.674651] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.674812] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.674988] env[62066]: DEBUG nova.virt.hardware [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.675869] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbda0818-7ae5-46f6-b81f-123629f2f8e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.683965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81bd24a-6766-43ed-9497-ac88686c11c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.849962] env[62066]: DEBUG nova.compute.manager [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Received event network-vif-plugged-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.850230] env[62066]: DEBUG oslo_concurrency.lockutils [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] Acquiring lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.850444] env[62066]: DEBUG oslo_concurrency.lockutils [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.850617] env[62066]: DEBUG oslo_concurrency.lockutils [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.850791] env[62066]: DEBUG nova.compute.manager [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] No waiting events found dispatching network-vif-plugged-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.850963] env[62066]: WARNING nova.compute.manager [req-4e3ecddd-6d5f-4ae3-bb15-54942a1ad6e8 req-315b65b2-53a9-4ee7-aa30-d1a732539618 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Received unexpected event network-vif-plugged-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d for instance with vm_state building and task_state spawning. [ 993.857485] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.881564] env[62066]: INFO nova.scheduler.client.report [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 7da117c1-0686-4b7e-9c33-5935a103dc3e [ 993.903832] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156633, 'name': PowerOffVM_Task, 'duration_secs': 0.182828} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.904183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.904396] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 993.904676] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1ce4009-b23a-491d-9b0b-5da517d093fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.956740] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Successfully updated port: d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.971188] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156634, 'name': ReconfigVM_Task, 'duration_secs': 0.203327} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.971188] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 993.972086] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66531db2-fe18-485c-8ca6-5ed213617f72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.998273] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.999061] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d985c1e0-9adf-4a17-9cde-a792dfe2b5e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.019468] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 994.019468] env[62066]: value = "task-1156636" [ 994.019468] env[62066]: _type = "Task" [ 994.019468] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.023723] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.023970] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.024183] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Deleting the datastore file [datastore2] 16390d89-f2d1-4d6c-be62-b85e61906865 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.024753] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef1240f8-8254-46b8-8497-5963764e65cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.031463] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156636, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.037214] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for the task: (returnval){ [ 994.037214] env[62066]: value = "task-1156637" [ 994.037214] env[62066]: _type = "Task" [ 994.037214] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.044659] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.061384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.061649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.061869] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.062155] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.062359] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.064671] env[62066]: INFO nova.compute.manager [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Terminating instance [ 994.066542] env[62066]: DEBUG nova.compute.manager [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 994.066742] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.067572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9a0e5c-9b33-458e-ad36-610887951a6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.074852] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.075125] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b842a3c-b337-4cfd-8a8c-687116a1ccba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.142830] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.143075] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.143257] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore2] 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.143524] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae596eec-3d8d-4820-8082-6505ac6bc90e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.149932] env[62066]: DEBUG oslo_vmware.api [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 994.149932] env[62066]: value = "task-1156639" [ 994.149932] env[62066]: _type = "Task" [ 994.149932] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.157674] env[62066]: DEBUG oslo_vmware.api [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.401565] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2a041e6-84ba-44d2-a5d2-3fd9b19c4b83 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "7da117c1-0686-4b7e-9c33-5935a103dc3e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.732s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.459608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.459785] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.460027] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.529790] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156636, 'name': ReconfigVM_Task, 'duration_secs': 0.360622} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.530078] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.530350] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 994.546334] env[62066]: DEBUG oslo_vmware.api [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Task: {'id': task-1156637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146812} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.546564] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.546746] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 994.546920] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 994.547110] env[62066]: INFO nova.compute.manager [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Took 1.19 seconds to destroy the instance on the hypervisor. [ 994.547421] env[62066]: DEBUG oslo.service.loopingcall [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.547629] env[62066]: DEBUG nova.compute.manager [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.547729] env[62066]: DEBUG nova.network.neutron [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.661340] env[62066]: DEBUG oslo_vmware.api [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176001} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.661556] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.661768] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 994.662047] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 994.662162] env[62066]: INFO nova.compute.manager [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Took 0.60 seconds to destroy the instance on the hypervisor. [ 994.662414] env[62066]: DEBUG oslo.service.loopingcall [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.662613] env[62066]: DEBUG nova.compute.manager [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.662706] env[62066]: DEBUG nova.network.neutron [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.992174] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 995.036407] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31685c1-1982-4219-9f3e-1f51a3694ed0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.061298] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f4c0b1-4c97-47c8-8a11-b3569107ade0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.083709] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 995.142593] env[62066]: DEBUG nova.network.neutron [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updating instance_info_cache with network_info: [{"id": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "address": "fa:16:3e:1d:84:ea", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ccc7dd-b3", "ovs_interfaceid": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.595225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.595225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.595398] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.595470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.595642] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.597906] env[62066]: INFO nova.compute.manager [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Terminating instance [ 995.603241] env[62066]: DEBUG nova.compute.manager [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.603241] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 995.603241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514526c8-6044-4a81-8a56-4abf588a3943 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.613371] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 995.614393] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce6eae15-9638-48f4-8db1-d81e689b9713 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.624535] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 995.624535] env[62066]: value = "task-1156640" [ 995.624535] env[62066]: _type = "Task" [ 995.624535] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.632280] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.644271] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.644937] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Instance network_info: |[{"id": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "address": "fa:16:3e:1d:84:ea", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ccc7dd-b3", "ovs_interfaceid": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.645745] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:84:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4ccc7dd-b3cd-42b7-9363-9424832a0a2d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.654616] env[62066]: DEBUG oslo.service.loopingcall [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.655262] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.659023] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2519112-4907-4ff2-b2b2-1731bda06638 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.679128] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.679128] env[62066]: value = "task-1156641" [ 995.679128] env[62066]: _type = "Task" [ 995.679128] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.687229] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156641, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.808375] env[62066]: DEBUG nova.network.neutron [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.886638] env[62066]: DEBUG nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Received event network-changed-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.886881] env[62066]: DEBUG nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Refreshing instance network info cache due to event network-changed-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.887108] env[62066]: DEBUG oslo_concurrency.lockutils [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] Acquiring lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.887264] env[62066]: DEBUG oslo_concurrency.lockutils [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] Acquired lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.887486] env[62066]: DEBUG nova.network.neutron [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Refreshing network info cache for port d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.934477] env[62066]: DEBUG nova.network.neutron [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.133400] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156640, 'name': PowerOffVM_Task, 'duration_secs': 0.335667} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.133765] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.134024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.134338] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d2af284-0cbb-4fa9-a647-1354ffd0bc4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.189981] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156641, 'name': CreateVM_Task, 'duration_secs': 0.474793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.190264] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.191561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.191766] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.192107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.192382] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61622820-ff0c-4000-a136-bb9ba48bf731 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.197117] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 996.197117] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffaa3b-f515-546b-6d39-3941e9fcb310" [ 996.197117] env[62066]: _type = "Task" [ 996.197117] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.206568] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffaa3b-f515-546b-6d39-3941e9fcb310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.208305] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.208508] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.208702] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleting the datastore file [datastore2] 212dac6a-a291-4ca8-87fb-97ebcca7976c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.208977] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a44c828-9a49-4f65-bc99-d09364b7b471 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.214777] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for the task: (returnval){ [ 996.214777] env[62066]: value = "task-1156643" [ 996.214777] env[62066]: _type = "Task" [ 996.214777] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.224672] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.311497] env[62066]: INFO nova.compute.manager [-] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Took 1.65 seconds to deallocate network for instance. [ 996.436981] env[62066]: INFO nova.compute.manager [-] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Took 1.89 seconds to deallocate network for instance. [ 996.602205] env[62066]: DEBUG nova.network.neutron [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updated VIF entry in instance network info cache for port d4ccc7dd-b3cd-42b7-9363-9424832a0a2d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.602594] env[62066]: DEBUG nova.network.neutron [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updating instance_info_cache with network_info: [{"id": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "address": "fa:16:3e:1d:84:ea", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ccc7dd-b3", "ovs_interfaceid": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.708639] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ffaa3b-f515-546b-6d39-3941e9fcb310, 'name': SearchDatastore_Task, 'duration_secs': 0.014563} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.708946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.709210] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.709455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.709605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.709786] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.710063] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c3d02ca-1535-4c63-a4ec-b0f9ed595943 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.719229] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.719398] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.720385] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d72bf857-1da3-4952-82c3-a9fd04494ca9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.727091] env[62066]: DEBUG oslo_vmware.api [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Task: {'id': task-1156643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191071} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.727631] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.727820] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.728009] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.728201] env[62066]: INFO nova.compute.manager [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 996.728425] env[62066]: DEBUG oslo.service.loopingcall [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.728897] env[62066]: DEBUG nova.compute.manager [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.729022] env[62066]: DEBUG nova.network.neutron [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 996.731409] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 996.731409] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5209ab05-5401-5b1b-36aa-1693880ecc78" [ 996.731409] env[62066]: _type = "Task" [ 996.731409] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.740305] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5209ab05-5401-5b1b-36aa-1693880ecc78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.789089] env[62066]: DEBUG nova.network.neutron [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 996.820704] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.821116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.821394] env[62066]: DEBUG nova.objects.instance [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'resources' on Instance uuid 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.946190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.105077] env[62066]: DEBUG oslo_concurrency.lockutils [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] Releasing lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.105419] env[62066]: DEBUG nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Received event network-vif-deleted-e1c0d760-32cf-44fb-b45e-0aa1b25c47ac {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.105638] env[62066]: DEBUG nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Received event network-vif-deleted-f414725f-8af3-4a19-acf5-6cc44ff37ecb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.105817] env[62066]: INFO nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Neutron deleted interface f414725f-8af3-4a19-acf5-6cc44ff37ecb; detaching it from the instance and deleting it from the info cache [ 997.105998] env[62066]: DEBUG nova.network.neutron [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.242657] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5209ab05-5401-5b1b-36aa-1693880ecc78, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.243459] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2ada409-13f3-4105-8bc2-b7d1025653cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.248835] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 997.248835] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528fac39-ac51-ac56-14d7-e5e5898427a8" [ 997.248835] env[62066]: _type = "Task" [ 997.248835] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.256394] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528fac39-ac51-ac56-14d7-e5e5898427a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.454956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0859860e-9abf-4a88-a07f-41e2a0d2c1e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.462445] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168ce162-c2ad-4800-b939-08221dd7a038 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.493135] env[62066]: DEBUG nova.network.neutron [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.494951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8352de21-79e1-4763-ab37-890b6e476e8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.502400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7804f8e4-8a8f-4b49-aa1b-5debc8ffeb0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.516340] env[62066]: DEBUG nova.compute.provider_tree [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.608858] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91c330b7-3eaa-4ecd-bf83-d2631c0ab98d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.621122] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60592451-8d27-47e9-9523-8e60afeb7e30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.646972] env[62066]: DEBUG nova.compute.manager [req-6e90d178-d609-4b16-af0b-5ee2b70291b1 req-b39ab3f7-00f1-450d-b0d0-832edc4b53e8 service nova] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Detach interface failed, port_id=f414725f-8af3-4a19-acf5-6cc44ff37ecb, reason: Instance 16390d89-f2d1-4d6c-be62-b85e61906865 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 997.759443] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528fac39-ac51-ac56-14d7-e5e5898427a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.759722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.760013] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6e76c799-9341-4315-8fa9-3c9acc1f8869/6e76c799-9341-4315-8fa9-3c9acc1f8869.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.760279] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7084a16d-485e-420a-be7c-8a8f3f4f4492 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.766468] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 997.766468] env[62066]: value = "task-1156644" [ 997.766468] env[62066]: _type = "Task" [ 997.766468] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.775037] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.810810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.810810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.810973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.913201] env[62066]: DEBUG nova.compute.manager [req-9ed79cba-77b1-4640-9ad9-bb99ae748525 req-51312fc0-bbfb-4a7b-b0b9-5a54ec74bd14 service nova] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Received event network-vif-deleted-59b6e38f-b186-420c-89c8-c860e8310108 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.999723] env[62066]: INFO nova.compute.manager [-] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Took 1.27 seconds to deallocate network for instance. [ 998.020057] env[62066]: DEBUG nova.scheduler.client.report [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.276334] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472124} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.276676] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 6e76c799-9341-4315-8fa9-3c9acc1f8869/6e76c799-9341-4315-8fa9-3c9acc1f8869.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.276847] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.277065] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-779b7c58-58a9-4e98-98bc-b9bab6f06701 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.283304] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 998.283304] env[62066]: value = "task-1156645" [ 998.283304] env[62066]: _type = "Task" [ 998.283304] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.290541] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.506530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.525679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.528025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.582s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.528202] env[62066]: DEBUG nova.objects.instance [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lazy-loading 'resources' on Instance uuid 16390d89-f2d1-4d6c-be62-b85e61906865 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.547084] env[62066]: INFO nova.scheduler.client.report [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc [ 998.793300] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063872} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.793667] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.794496] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc88a68f-0db2-4e7d-9cc8-5d84200f57b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.815852] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 6e76c799-9341-4315-8fa9-3c9acc1f8869/6e76c799-9341-4315-8fa9-3c9acc1f8869.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.818665] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2c70a87-5102-47e6-bd27-cddd544b2ddb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.838589] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 998.838589] env[62066]: value = "task-1156646" [ 998.838589] env[62066]: _type = "Task" [ 998.838589] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.846674] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.862990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.863213] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.863395] env[62066]: DEBUG nova.network.neutron [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.054389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69e33660-de99-4f09-96d6-d55d5388d608 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.993s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.160046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031fab97-5203-4fcf-9ef0-07dd47eb0531 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.167670] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bce9e3-70d2-4d48-8e0c-5c49a4c1c148 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.196679] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d15bc63-f367-446a-9429-6383ec1db28c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.203628] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4faa7e-e60d-4a55-9d92-263bd318891c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.216512] env[62066]: DEBUG nova.compute.provider_tree [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.348373] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156646, 'name': ReconfigVM_Task, 'duration_secs': 0.270414} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.348655] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 6e76c799-9341-4315-8fa9-3c9acc1f8869/6e76c799-9341-4315-8fa9-3c9acc1f8869.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.349301] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d781df06-f678-4702-b640-43ac216954e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.355471] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 999.355471] env[62066]: value = "task-1156647" [ 999.355471] env[62066]: _type = "Task" [ 999.355471] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.363379] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156647, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.590188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.590509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.592409] env[62066]: DEBUG nova.objects.instance [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid 6e0065b0-da4f-4288-a858-8648572f8148 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.598616] env[62066]: DEBUG nova.network.neutron [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.719814] env[62066]: DEBUG nova.scheduler.client.report [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.865682] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156647, 'name': Rename_Task, 'duration_secs': 0.142843} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.865963] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.866213] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-746733b9-35a4-47af-9ffd-e45faa571f80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.872504] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 999.872504] env[62066]: value = "task-1156648" [ 999.872504] env[62066]: _type = "Task" [ 999.872504] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.879637] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156648, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.081598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.081806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.102977] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.189052] env[62066]: DEBUG nova.objects.instance [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid 6e0065b0-da4f-4288-a858-8648572f8148 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.224381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.226772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.720s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.227024] env[62066]: DEBUG nova.objects.instance [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lazy-loading 'resources' on Instance uuid 212dac6a-a291-4ca8-87fb-97ebcca7976c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.242939] env[62066]: INFO nova.scheduler.client.report [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Deleted allocations for instance 16390d89-f2d1-4d6c-be62-b85e61906865 [ 1000.382857] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156648, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.584319] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.613876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cbfbe9-23ef-4d45-b679-1933089b1887 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.621667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218cf5a9-ec09-43ed-9285-0c5f3990b45d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.691398] env[62066]: DEBUG nova.objects.base [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance<6e0065b0-da4f-4288-a858-8648572f8148> lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1000.691640] env[62066]: DEBUG nova.network.neutron [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1000.754219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fb16932d-48dd-4826-ac72-c057357c93e5 tempest-ServersTestFqdnHostnames-1209090242 tempest-ServersTestFqdnHostnames-1209090242-project-member] Lock "16390d89-f2d1-4d6c-be62-b85e61906865" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.400s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.755765] env[62066]: DEBUG nova.policy [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1000.852473] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd5f4ad-45a8-469e-b0a4-4d3077958306 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.859974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444ca53c-36f0-494e-9789-8c445b578714 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.896017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fa5ab9-640c-42a7-8461-f7114f698c48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.903837] env[62066]: DEBUG oslo_vmware.api [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156648, 'name': PowerOnVM_Task, 'duration_secs': 0.52835} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.905943] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.906177] env[62066]: INFO nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1000.906417] env[62066]: DEBUG nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.907650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4309c3-5239-4475-b730-005f6a038ac4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.911218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e173d6d2-9e5a-4dd9-a991-3f60e151ea8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.928869] env[62066]: DEBUG nova.compute.provider_tree [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.067342] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.067621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.067928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.068048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.068338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.070574] env[62066]: INFO nova.compute.manager [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Terminating instance [ 1001.072341] env[62066]: DEBUG nova.compute.manager [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.072536] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.073390] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d6ea7e-7d7e-4b47-ac27-b93b02c309fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.081085] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.081085] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b24a3eaa-85f7-4650-9a4b-b66cd5e448cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.086723] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1001.086723] env[62066]: value = "task-1156649" [ 1001.086723] env[62066]: _type = "Task" [ 1001.086723] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.098341] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.107436] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.435327] env[62066]: DEBUG nova.scheduler.client.report [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.442780] env[62066]: INFO nova.compute.manager [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Took 12.25 seconds to build instance. [ 1001.597208] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156649, 'name': PowerOffVM_Task, 'duration_secs': 0.175428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.597509] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.597704] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.597929] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fcd0fe9-6e93-42e5-802b-68734fd0b9de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.668628] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.669699] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.669699] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleting the datastore file [datastore2] 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.669699] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-965eabe5-25ac-4aa3-822a-e421faa0c661 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.675667] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1001.675667] env[62066]: value = "task-1156651" [ 1001.675667] env[62066]: _type = "Task" [ 1001.675667] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.683552] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.724544] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0960181-45cf-4102-ae56-87e9ba2b1781 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.746878] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b21056c-fe8e-4741-944e-e0e204dcdb90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.754448] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.942875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.945312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.838s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.946816] env[62066]: INFO nova.compute.claims [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.949322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6fb2c95-f1fc-45e9-87e3-01c83e0a2a7d tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.769s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.961558] env[62066]: INFO nova.scheduler.client.report [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Deleted allocations for instance 212dac6a-a291-4ca8-87fb-97ebcca7976c [ 1002.186501] env[62066]: DEBUG oslo_vmware.api [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171673} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.186851] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.187117] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.187370] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.187610] env[62066]: INFO nova.compute.manager [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1002.187913] env[62066]: DEBUG oslo.service.loopingcall [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.188182] env[62066]: DEBUG nova.compute.manager [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.188335] env[62066]: DEBUG nova.network.neutron [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1002.261246] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.261609] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5693e8a5-be3e-45de-be9e-77079bf0366d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.269850] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1002.269850] env[62066]: value = "task-1156652" [ 1002.269850] env[62066]: _type = "Task" [ 1002.269850] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.278358] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.469356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ce41c05f-fdd3-40f3-a643-d41e9c1ba8b0 tempest-ServersTestJSON-1214729601 tempest-ServersTestJSON-1214729601-project-member] Lock "212dac6a-a291-4ca8-87fb-97ebcca7976c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.874s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.477229] env[62066]: DEBUG nova.compute.manager [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Received event network-changed-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.477229] env[62066]: DEBUG nova.compute.manager [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Refreshing instance network info cache due to event network-changed-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.477229] env[62066]: DEBUG oslo_concurrency.lockutils [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] Acquiring lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.477229] env[62066]: DEBUG oslo_concurrency.lockutils [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] Acquired lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.477229] env[62066]: DEBUG nova.network.neutron [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Refreshing network info cache for port d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1002.704706] env[62066]: DEBUG nova.network.neutron [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Successfully updated port: 27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.740350] env[62066]: DEBUG nova.compute.manager [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.742747] env[62066]: DEBUG oslo_concurrency.lockutils [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.742747] env[62066]: DEBUG oslo_concurrency.lockutils [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.742747] env[62066]: DEBUG oslo_concurrency.lockutils [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.742747] env[62066]: DEBUG nova.compute.manager [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] No waiting events found dispatching network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.742747] env[62066]: WARNING nova.compute.manager [req-db3a6beb-85db-475c-8732-69c4fd0e15c0 req-f2382f63-f0a0-4284-9963-02f1cf9a7a9b service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received unexpected event network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 for instance with vm_state active and task_state None. [ 1002.781670] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156652, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.102128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babc13db-c4c4-4890-aa94-61c48b463200 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.109950] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdadb9c7-8d31-4ce5-86ce-decc6d8bad98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.154447] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3e12b8-e29a-4012-8233-cb778b5465e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.164205] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcd0762-c7d9-42f7-9bb6-1ba04005032a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.177983] env[62066]: DEBUG nova.compute.provider_tree [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1003.211679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.211870] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.212183] env[62066]: DEBUG nova.network.neutron [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.282205] env[62066]: DEBUG oslo_vmware.api [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156652, 'name': PowerOnVM_Task, 'duration_secs': 0.644822} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.282561] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.282785] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5babf84d-fa87-4c9c-8252-e0592a68c40a tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.290520] env[62066]: DEBUG nova.network.neutron [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updated VIF entry in instance network info cache for port d4ccc7dd-b3cd-42b7-9363-9424832a0a2d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1003.290520] env[62066]: DEBUG nova.network.neutron [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updating instance_info_cache with network_info: [{"id": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "address": "fa:16:3e:1d:84:ea", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4ccc7dd-b3", "ovs_interfaceid": "d4ccc7dd-b3cd-42b7-9363-9424832a0a2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.389572] env[62066]: DEBUG nova.network.neutron [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.700304] env[62066]: ERROR nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [req-a84b5e36-2275-4af0-b34a-70398fc470c9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a84b5e36-2275-4af0-b34a-70398fc470c9"}]} [ 1003.724190] env[62066]: DEBUG nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1003.745959] env[62066]: DEBUG nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1003.746307] env[62066]: DEBUG nova.compute.provider_tree [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1003.758604] env[62066]: WARNING nova.network.neutron [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 1003.765214] env[62066]: DEBUG nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1003.787049] env[62066]: DEBUG nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1003.792956] env[62066]: DEBUG oslo_concurrency.lockutils [req-43340c8a-c0ec-49ef-b234-0ae545769931 req-89553e43-4e8a-4934-8342-cbc479fb7751 service nova] Releasing lock "refresh_cache-6e76c799-9341-4315-8fa9-3c9acc1f8869" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.894439] env[62066]: INFO nova.compute.manager [-] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Took 1.71 seconds to deallocate network for instance. [ 1003.946923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4e2b97-78a1-434e-84a3-618e436527ff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.955677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4905a79f-7938-43a4-8575-42501d72bcaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.988251] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33df21a1-d390-47a4-a8fd-8a1873f05504 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.999914] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae92066-e941-4bef-a517-52cadba1b2ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.019017] env[62066]: DEBUG nova.compute.provider_tree [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.401785] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.478713] env[62066]: DEBUG nova.network.neutron [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "address": "fa:16:3e:06:2e:5f", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27157a0e-62", "ovs_interfaceid": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.568339] env[62066]: DEBUG nova.scheduler.client.report [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1004.569726] env[62066]: DEBUG nova.compute.provider_tree [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 131 to 132 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1004.569726] env[62066]: DEBUG nova.compute.provider_tree [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.776164] env[62066]: DEBUG nova.compute.manager [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.776477] env[62066]: DEBUG nova.compute.manager [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-27157a0e-629d-45bb-9bdf-b8e235ce8be0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.776611] env[62066]: DEBUG oslo_concurrency.lockutils [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.982792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.983549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.983720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.984018] env[62066]: DEBUG oslo_concurrency.lockutils [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.984299] env[62066]: DEBUG nova.network.neutron [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.986151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b9a8fd-0c88-4008-9c6e-0334a0960bc4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.005778] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.006040] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.006209] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.006402] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.006625] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.006785] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.006997] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.007181] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.007355] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.007523] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.007704] env[62066]: DEBUG nova.virt.hardware [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.013934] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1005.014988] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4960559b-c14c-4616-905d-6d90b018ee50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.038761] env[62066]: DEBUG oslo_vmware.api [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1005.038761] env[62066]: value = "task-1156654" [ 1005.038761] env[62066]: _type = "Task" [ 1005.038761] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.048199] env[62066]: DEBUG oslo_vmware.api [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156654, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.077334] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.132s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.077985] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.081505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.680s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.081768] env[62066]: DEBUG nova.objects.instance [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'resources' on Instance uuid 31e50f97-f873-44ad-9923-67923cdb8d3a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.550166] env[62066]: DEBUG oslo_vmware.api [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.582909] env[62066]: DEBUG nova.compute.utils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1005.588024] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.588024] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.704018] env[62066]: DEBUG nova.policy [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.737903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bb36d8-fe5b-4e4e-b7d7-411417505f1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.748126] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7bcd52-71d8-4825-b665-4c0103849f6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.786236] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b679a42-2719-474b-a839-f281ce93a201 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.797983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2d5da9-457e-4711-9798-ede7a81dd126 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.815692] env[62066]: DEBUG nova.compute.provider_tree [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.870671] env[62066]: DEBUG nova.network.neutron [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 27157a0e-629d-45bb-9bdf-b8e235ce8be0. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1005.871139] env[62066]: DEBUG nova.network.neutron [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "address": "fa:16:3e:06:2e:5f", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27157a0e-62", "ovs_interfaceid": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.050218] env[62066]: DEBUG oslo_vmware.api [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156654, 'name': ReconfigVM_Task, 'duration_secs': 0.798845} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.050761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.051008] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1006.091991] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.098125] env[62066]: DEBUG nova.network.neutron [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Port f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1006.098125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.098125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.098125] env[62066]: DEBUG nova.network.neutron [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1006.207797] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Successfully created port: f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.321144] env[62066]: DEBUG nova.scheduler.client.report [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.373522] env[62066]: DEBUG oslo_concurrency.lockutils [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.373876] env[62066]: DEBUG nova.compute.manager [req-788fcaec-d50f-48dd-b0d9-7206a8db6f83 req-667c4e7f-b042-42f6-9adf-a96383079da4 service nova] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Received event network-vif-deleted-6244eb3b-c073-494b-b711-f8712351da75 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.556574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d351b9b-c6dc-4f5b-a193-967ae625a1be tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.966s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.698217] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.829643] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.882652] env[62066]: INFO nova.scheduler.client.report [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleted allocations for instance 31e50f97-f873-44ad-9923-67923cdb8d3a [ 1006.910950] env[62066]: DEBUG nova.network.neutron [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.106659] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.140686] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.141130] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.141525] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.141823] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.142117] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.142394] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.142704] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.143028] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.143352] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.144114] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.145062] env[62066]: DEBUG nova.virt.hardware [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.145971] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d915bf54-fae8-49c8-a4e9-13fad1f2a17c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.156423] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d601da2d-694d-497d-86ee-fc9ca740a33f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.201369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.201506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.201670] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.202213] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1007.202789] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ffa2d6-36be-4f01-a83b-60e4c578f277 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.214020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb064cd7-393b-4551-bb4a-b3e1088f2e36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.231159] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813e0727-5660-47d8-934f-7f112dcf9491 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.239051] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f89092-c89b-4b32-b73c-42cb12c91731 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.280162] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180224MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1007.280297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.280996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.396259] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3b85c5-88b3-4120-bf8c-7b829de6717a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "31e50f97-f873-44ad-9923-67923cdb8d3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.329s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.416157] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.700033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "e273b275-2c03-4ac6-919e-46fddde91903" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.701392] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.761233] env[62066]: DEBUG nova.compute.manager [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Received event network-vif-plugged-f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.761646] env[62066]: DEBUG oslo_concurrency.lockutils [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] Acquiring lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.762014] env[62066]: DEBUG oslo_concurrency.lockutils [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.762170] env[62066]: DEBUG oslo_concurrency.lockutils [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.762311] env[62066]: DEBUG nova.compute.manager [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] No waiting events found dispatching network-vif-plugged-f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.762746] env[62066]: WARNING nova.compute.manager [req-f7d82c42-5ae9-41e2-87f1-5b589d92f2ec req-1c532f1e-e51d-43ff-bf48-de04a3981d5f service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Received unexpected event network-vif-plugged-f51f939f-f9be-457d-8222-b65f88a1490f for instance with vm_state building and task_state spawning. [ 1007.857690] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Successfully updated port: f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.920141] env[62066]: DEBUG nova.compute.manager [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62066) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1007.948147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.948470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.206019] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.292454] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Applying migration context for instance ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 as it has an incoming, in-progress migration 02ce68b7-a528-4678-aa25-7fe4e6bfb631. Migration status is reverting {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1008.294641] env[62066]: INFO nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating resource usage from migration 02ce68b7-a528-4678-aa25-7fe4e6bfb631 [ 1008.330793] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.331280] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6e0065b0-da4f-4288-a858-8648572f8148 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.331693] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.331693] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Migration 02ce68b7-a528-4678-aa25-7fe4e6bfb631 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1008.331904] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.332185] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6e76c799-9341-4315-8fa9-3c9acc1f8869 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.332366] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 08bc5e66-939c-436e-8b7b-dff7f8ff50f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.360793] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.360992] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.361284] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.454796] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.455050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.456113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e1551c-fc89-48cc-93a6-bf2f6f0be3f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.481062] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-018128a6-dacf-4a8e-a682-06f4357e37b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.514517] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1008.514517] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-988094a2-cfc5-4cfe-a028-da59822d0cd5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.536552] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1008.536552] env[62066]: value = "task-1156656" [ 1008.536552] env[62066]: _type = "Task" [ 1008.536552] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.547636] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.735433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.837536] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance e273b275-2c03-4ac6-919e-46fddde91903 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1008.837722] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1008.837836] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1008.903041] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1009.018277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ccb203-7170-4717-b620-69373276b775 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.028629] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068c181b-b086-40f4-9342-3642817109b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.072056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.073195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc4cc96-a508-4e85-b4dc-f941ebf748b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.079439] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.085790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038f8ded-88c4-4ba9-a505-ef3dfba1d941 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.090590] env[62066]: DEBUG nova.network.neutron [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Updating instance_info_cache with network_info: [{"id": "f51f939f-f9be-457d-8222-b65f88a1490f", "address": "fa:16:3e:89:6d:a4", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf51f939f-f9", "ovs_interfaceid": "f51f939f-f9be-457d-8222-b65f88a1490f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.103946] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.549486] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.596363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.596363] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Instance network_info: |[{"id": "f51f939f-f9be-457d-8222-b65f88a1490f", "address": "fa:16:3e:89:6d:a4", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf51f939f-f9", "ovs_interfaceid": "f51f939f-f9be-457d-8222-b65f88a1490f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.596686] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:6d:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f51f939f-f9be-457d-8222-b65f88a1490f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.605778] env[62066]: DEBUG oslo.service.loopingcall [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.605869] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.606177] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78c0e778-52cb-437a-ba76-37a373dd6b27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.631640] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.631640] env[62066]: value = "task-1156658" [ 1009.631640] env[62066]: _type = "Task" [ 1009.631640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.641647] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156658, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.644605] env[62066]: ERROR nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [req-84d53cfd-12fe-491e-b1ce-1840bf58228f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-84d53cfd-12fe-491e-b1ce-1840bf58228f"}]} [ 1009.665966] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1009.687972] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1009.687972] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.705614] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1009.725979] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1009.783806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.783936] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.789019] env[62066]: DEBUG nova.compute.manager [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Received event network-changed-f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.790523] env[62066]: DEBUG nova.compute.manager [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Refreshing instance network info cache due to event network-changed-f51f939f-f9be-457d-8222-b65f88a1490f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.790523] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] Acquiring lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.790523] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] Acquired lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.790523] env[62066]: DEBUG nova.network.neutron [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Refreshing network info cache for port f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.862902] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e2aed5-346b-4451-84f8-2b2dc3c649fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.874480] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630398c8-5106-49e9-9f08-ea820dfe4968 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.916065] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b13979-e79d-48b6-8b84-1c878551fd49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.924953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0ea023-8a97-4bb6-bd2a-c9fe4d00ec06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.941540] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.049745] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.142650] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156658, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.289201] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.495071] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1010.499286] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 134 to 135 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1010.499286] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.504807] env[62066]: DEBUG nova.network.neutron [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Updated VIF entry in instance network info cache for port f51f939f-f9be-457d-8222-b65f88a1490f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.504980] env[62066]: DEBUG nova.network.neutron [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Updating instance_info_cache with network_info: [{"id": "f51f939f-f9be-457d-8222-b65f88a1490f", "address": "fa:16:3e:89:6d:a4", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf51f939f-f9", "ovs_interfaceid": "f51f939f-f9be-457d-8222-b65f88a1490f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.550374] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.644071] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156658, 'name': CreateVM_Task, 'duration_secs': 0.672196} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.644071] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1010.644916] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.644984] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.645299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.645566] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bec65c4-4610-44d6-86db-26dbfbc394fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.652274] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1010.652274] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bcb413-b9d8-5d52-2ae6-e0833983daaa" [ 1010.652274] env[62066]: _type = "Task" [ 1010.652274] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.661167] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bcb413-b9d8-5d52-2ae6-e0833983daaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.809649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.001960] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1011.002874] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.722s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.003252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.268s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.004838] env[62066]: INFO nova.compute.claims [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.007710] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac1dae24-2b5d-4708-a875-1d66f59af795 req-0c75461d-63e6-4fef-8e94-418bf213f394 service nova] Releasing lock "refresh_cache-08bc5e66-939c-436e-8b7b-dff7f8ff50f4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.053499] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.134720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.134948] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.166498] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bcb413-b9d8-5d52-2ae6-e0833983daaa, 'name': SearchDatastore_Task, 'duration_secs': 0.011131} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.166830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.167094] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.167358] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.167531] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.167729] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.168450] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82ab83db-cf12-490a-93df-66d9b0503358 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.178486] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.178632] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.179712] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff64291-1d74-47dd-9348-06bf55b20e86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.185613] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1011.185613] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bbe233-4942-18ce-76ce-6ae1d35fbc47" [ 1011.185613] env[62066]: _type = "Task" [ 1011.185613] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.193945] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bbe233-4942-18ce-76ce-6ae1d35fbc47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.551794] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.637165] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.696430] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bbe233-4942-18ce-76ce-6ae1d35fbc47, 'name': SearchDatastore_Task, 'duration_secs': 0.009054} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.697241] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94ae762b-9a25-46c2-a5f6-98d99db12085 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.702713] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1011.702713] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d4fd42-1559-8795-5621-200ad35cb3e1" [ 1011.702713] env[62066]: _type = "Task" [ 1011.702713] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.711290] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d4fd42-1559-8795-5621-200ad35cb3e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.057346] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.146851] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75351f0b-5511-4cae-b329-31fb60a7a8ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.154582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.155655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461baeb9-a665-4566-8d62-198ed4a671d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.188604] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7e96c3-e654-4b22-a395-c5308b7884de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.196888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cdad88-a5d6-4924-a189-3dadfbe83205 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.908415] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d4fd42-1559-8795-5621-200ad35cb3e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009271} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.918503] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.918787] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 08bc5e66-939c-436e-8b7b-dff7f8ff50f4/08bc5e66-939c-436e-8b7b-dff7f8ff50f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.919279] env[62066]: DEBUG nova.compute.provider_tree [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.920421] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.920806] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81832ccc-b7a2-4bf3-a942-3e1b179ed0a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.930070] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1012.930070] env[62066]: value = "task-1156660" [ 1012.930070] env[62066]: _type = "Task" [ 1012.930070] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.938019] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.408787] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.422009] env[62066]: DEBUG nova.scheduler.client.report [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.441350] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499893} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.441350] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 08bc5e66-939c-436e-8b7b-dff7f8ff50f4/08bc5e66-939c-436e-8b7b-dff7f8ff50f4.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.441350] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.441350] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c396f559-8cdc-4a98-9b30-828283584f77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.451609] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1013.451609] env[62066]: value = "task-1156662" [ 1013.451609] env[62066]: _type = "Task" [ 1013.451609] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.460893] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.908855] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.926955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.924s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.927466] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.930181] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 4.859s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.962186] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0689} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.962550] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.963343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18215e48-8ab9-491d-b5cc-f9ce239a850a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.987467] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 08bc5e66-939c-436e-8b7b-dff7f8ff50f4/08bc5e66-939c-436e-8b7b-dff7f8ff50f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.988107] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c74493a-08d6-4cd9-b1b1-3dc2e26762ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.009355] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1014.010919] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1014.011093] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1014.012487] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1014.012487] env[62066]: value = "task-1156663" [ 1014.012487] env[62066]: _type = "Task" [ 1014.012487] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.022787] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156663, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.408717] env[62066]: DEBUG oslo_vmware.api [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156656, 'name': ReconfigVM_Task, 'duration_secs': 5.836872} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.409011] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.409272] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1014.433438] env[62066]: DEBUG nova.compute.utils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1014.435164] env[62066]: DEBUG nova.objects.instance [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'migration_context' on Instance uuid ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.436422] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.436592] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.505317] env[62066]: DEBUG nova.policy [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '71133269ae0c4d778f5164ec4311fe23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e62ef3f429234c38831e496abac10869', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.529749] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156663, 'name': ReconfigVM_Task, 'duration_secs': 0.371694} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.531201] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 08bc5e66-939c-436e-8b7b-dff7f8ff50f4/08bc5e66-939c-436e-8b7b-dff7f8ff50f4.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.531757] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b7224d4-106c-462b-b606-1290e3923ee2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.539053] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1014.539053] env[62066]: value = "task-1156664" [ 1014.539053] env[62066]: _type = "Task" [ 1014.539053] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.547974] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156664, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.548745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.548876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.549030] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1014.937910] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.964080] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Successfully created port: 256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.055380] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156664, 'name': Rename_Task, 'duration_secs': 0.145891} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.055686] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.056405] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a767bff7-4988-4570-baba-0ccb9c3131c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.067043] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1015.067043] env[62066]: value = "task-1156665" [ 1015.067043] env[62066]: _type = "Task" [ 1015.067043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.079783] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156665, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.130760] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a59bf6-aa22-451d-9d76-090930ebc21e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.140552] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5eb6024-da06-436b-a976-709930364d66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.176386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087afb79-5d1b-4311-931a-4cc7b067ac74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.186552] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f653d5f-b69c-4279-ac73-bab4b4801ff5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.202299] env[62066]: DEBUG nova.compute.provider_tree [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.579126] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156665, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.705366] env[62066]: DEBUG nova.scheduler.client.report [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.804882] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.952142] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.955372] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.955621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.956014] env[62066]: DEBUG nova.network.neutron [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1015.978545] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.978838] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.978991] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.979251] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.979422] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.979570] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.979777] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.979939] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.980128] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.980296] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.980471] env[62066]: DEBUG nova.virt.hardware [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.981379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c70b0b-7dcf-4003-8629-19e0d889f977 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.990234] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ff6fdf-4f67-44e3-983e-0c752682c4e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.079600] env[62066]: DEBUG oslo_vmware.api [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156665, 'name': PowerOnVM_Task, 'duration_secs': 0.854471} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.080083] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.080176] env[62066]: INFO nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Took 8.97 seconds to spawn the instance on the hypervisor. [ 1016.080360] env[62066]: DEBUG nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.081157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f809462-e658-4a3d-b1d5-4a4662bce059 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.309536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.309536] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1016.309536] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1016.432030] env[62066]: DEBUG nova.compute.manager [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Received event network-vif-plugged-256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.432285] env[62066]: DEBUG oslo_concurrency.lockutils [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] Acquiring lock "e273b275-2c03-4ac6-919e-46fddde91903-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.432501] env[62066]: DEBUG oslo_concurrency.lockutils [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] Lock "e273b275-2c03-4ac6-919e-46fddde91903-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.432674] env[62066]: DEBUG oslo_concurrency.lockutils [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] Lock "e273b275-2c03-4ac6-919e-46fddde91903-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.432842] env[62066]: DEBUG nova.compute.manager [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] No waiting events found dispatching network-vif-plugged-256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.433014] env[62066]: WARNING nova.compute.manager [req-303421ab-4af8-42f6-aaed-fc3ad9c65447 req-6ba419f4-2358-451d-85e8-df01ddebc31b service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Received unexpected event network-vif-plugged-256b055b-e5b6-4f7d-97c7-d21c01586718 for instance with vm_state building and task_state spawning. [ 1016.534450] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Successfully updated port: 256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.600736] env[62066]: INFO nova.compute.manager [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Took 15.51 seconds to build instance. [ 1016.694749] env[62066]: INFO nova.network.neutron [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Port 27157a0e-629d-45bb-9bdf-b8e235ce8be0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1016.695192] env[62066]: DEBUG nova.network.neutron [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.716760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.786s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.723662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.914s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.725335] env[62066]: INFO nova.compute.claims [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.851563] env[62066]: DEBUG nova.compute.manager [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.851854] env[62066]: DEBUG nova.compute.manager [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing instance network info cache due to event network-changed-00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.851980] env[62066]: DEBUG oslo_concurrency.lockutils [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] Acquiring lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.037056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.037218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquired lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.037372] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.102698] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c0c50412-0073-407a-a53d-d0e26f9811ed tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.021s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.198278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.200948] env[62066]: DEBUG oslo_concurrency.lockutils [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] Acquired lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.201275] env[62066]: DEBUG nova.network.neutron [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Refreshing network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.375302] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e59960-4bf9-480f-b8fd-295f1308e1d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.383099] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1017.383381] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-661bede8-a3b3-4050-8534-0c59e8694f67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.390668] env[62066]: DEBUG oslo_vmware.api [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1017.390668] env[62066]: value = "task-1156668" [ 1017.390668] env[62066]: _type = "Task" [ 1017.390668] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.398864] env[62066]: DEBUG oslo_vmware.api [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156668, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.543638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.543944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.544361] env[62066]: DEBUG nova.objects.instance [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'flavor' on Instance uuid 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.574085] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.701963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f272357d-6a0a-4e16-a09f-3596c0bc28ea tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-6e0065b0-da4f-4288-a858-8648572f8148-27157a0e-629d-45bb-9bdf-b8e235ce8be0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.753s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.752180] env[62066]: DEBUG nova.network.neutron [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Updating instance_info_cache with network_info: [{"id": "256b055b-e5b6-4f7d-97c7-d21c01586718", "address": "fa:16:3e:68:17:b4", "network": {"id": "106cfd91-6d32-4e17-8e5d-a147084c0206", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1997972038-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e62ef3f429234c38831e496abac10869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap256b055b-e5", "ovs_interfaceid": "256b055b-e5b6-4f7d-97c7-d21c01586718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.890158] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7601bd7e-2f9e-4ec3-8a4e-eae04b462a78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.907494] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f299014-04cb-4274-9f9e-70987598eade {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.911593] env[62066]: DEBUG oslo_vmware.api [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156668, 'name': SuspendVM_Task} progress is 62%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.941810] env[62066]: DEBUG nova.network.neutron [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updated VIF entry in instance network info cache for port 00ffbc69-586f-4442-b0fc-616ec3c6c49b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.942518] env[62066]: DEBUG nova.network.neutron [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [{"id": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "address": "fa:16:3e:2c:aa:fb", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00ffbc69-58", "ovs_interfaceid": "00ffbc69-586f-4442-b0fc-616ec3c6c49b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.944688] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca969f4-b516-4965-8c97-8d4c29ce59a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.955033] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e626270-3c81-4825-800b-d7448a4fd7e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.973665] env[62066]: DEBUG nova.compute.provider_tree [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.141799] env[62066]: DEBUG nova.objects.instance [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'pci_requests' on Instance uuid 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.257222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Releasing lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.257556] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Instance network_info: |[{"id": "256b055b-e5b6-4f7d-97c7-d21c01586718", "address": "fa:16:3e:68:17:b4", "network": {"id": "106cfd91-6d32-4e17-8e5d-a147084c0206", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1997972038-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e62ef3f429234c38831e496abac10869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap256b055b-e5", "ovs_interfaceid": "256b055b-e5b6-4f7d-97c7-d21c01586718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.258213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:17:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd0bdd11b-58af-4cc0-9d38-8322e1bb4e74', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '256b055b-e5b6-4f7d-97c7-d21c01586718', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.265838] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Creating folder: Project (e62ef3f429234c38831e496abac10869). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1018.266290] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d11ad3e-a825-447e-a869-735a34192e2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.269099] env[62066]: INFO nova.compute.manager [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Swapping old allocation on dict_keys(['cd4c0e36-9c88-4f73-a93c-1ff383ed97c4']) held by migration 02ce68b7-a528-4678-aa25-7fe4e6bfb631 for instance [ 1018.282029] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Created folder: Project (e62ef3f429234c38831e496abac10869) in parent group-v251573. [ 1018.282343] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Creating folder: Instances. Parent ref: group-v251770. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1018.285109] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4b95531-04e2-47d3-8b31-ae6b62262c4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.296044] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Created folder: Instances in parent group-v251770. [ 1018.296334] env[62066]: DEBUG oslo.service.loopingcall [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.297213] env[62066]: DEBUG nova.scheduler.client.report [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Overwriting current allocation {'allocations': {'cd4c0e36-9c88-4f73-a93c-1ff383ed97c4': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 136}}, 'project_id': 'c468d5ba348d437f97a74e0da70bb42e', 'user_id': 'c41065776f39438a8f6d427c0f62b93f', 'consumer_generation': 1} on consumer ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 {{(pid=62066) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1018.298987] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1018.299405] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61e45ab4-65b9-462d-99eb-d0f4ac3843ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.318201] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.318201] env[62066]: value = "task-1156671" [ 1018.318201] env[62066]: _type = "Task" [ 1018.318201] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.325956] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156671, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.384776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.384964] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.385198] env[62066]: DEBUG nova.network.neutron [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.405598] env[62066]: DEBUG oslo_vmware.api [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156668, 'name': SuspendVM_Task, 'duration_secs': 0.691434} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.406012] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1018.406250] env[62066]: DEBUG nova.compute.manager [None req-ed30a760-ec4b-4993-b380-13be46e2c13e tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.407253] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badbbeb7-ac1f-45ae-a1cf-694c1348a8fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.450817] env[62066]: DEBUG oslo_concurrency.lockutils [req-75c83217-660d-4eca-a55e-0514094c1dbc req-1b27f19a-85fd-4675-bcdf-728c0b281849 service nova] Releasing lock "refresh_cache-6e0065b0-da4f-4288-a858-8648572f8148" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.493287] env[62066]: ERROR nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [req-3410cb39-4daa-463e-a9f6-2b4edf54bdc4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3410cb39-4daa-463e-a9f6-2b4edf54bdc4"}]} [ 1018.509356] env[62066]: DEBUG nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1018.522472] env[62066]: DEBUG nova.compute.manager [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Received event network-changed-256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.522691] env[62066]: DEBUG nova.compute.manager [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Refreshing instance network info cache due to event network-changed-256b055b-e5b6-4f7d-97c7-d21c01586718. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.522903] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] Acquiring lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.523453] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] Acquired lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.523705] env[62066]: DEBUG nova.network.neutron [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Refreshing network info cache for port 256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.526092] env[62066]: DEBUG nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1018.526295] env[62066]: DEBUG nova.compute.provider_tree [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.537866] env[62066]: DEBUG nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1018.556924] env[62066]: DEBUG nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1018.644340] env[62066]: DEBUG nova.objects.base [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Object Instance<045c2efa-e2fe-4d51-b267-e2b380fcf1d2> lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1018.644614] env[62066]: DEBUG nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.676664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456ce476-22d1-4224-bc7e-0593695ab860 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.685107] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6820b4e4-e53a-4e88-985f-1439d52d28d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.721629] env[62066]: DEBUG nova.policy [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95debd9e3bd9470ca0052f8bf0b19d83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42219a58a1514265b9d0b515eb517933', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1018.724167] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ba624b-d476-42ff-ad71-6c41d816a944 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.732117] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5096afc7-5ecb-4b25-92d0-f7d02659c951 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.745774] env[62066]: DEBUG nova.compute.provider_tree [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.828376] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156671, 'name': CreateVM_Task, 'duration_secs': 0.409626} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.828518] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.829223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.829397] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.829724] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.829973] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c39bbed-ecee-448d-9a42-bfdb73faca19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.835126] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1018.835126] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]527e54e9-2620-f1b3-643e-ea09566ba590" [ 1018.835126] env[62066]: _type = "Task" [ 1018.835126] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.842133] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527e54e9-2620-f1b3-643e-ea09566ba590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.876942] env[62066]: DEBUG nova.compute.manager [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.876942] env[62066]: DEBUG nova.compute.manager [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing instance network info cache due to event network-changed-567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.876942] env[62066]: DEBUG oslo_concurrency.lockutils [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.876942] env[62066]: DEBUG oslo_concurrency.lockutils [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.877826] env[62066]: DEBUG nova.network.neutron [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.137384] env[62066]: DEBUG nova.network.neutron [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [{"id": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "address": "fa:16:3e:d8:a3:01", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3c50a8a-9b", "ovs_interfaceid": "f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.271016] env[62066]: DEBUG nova.network.neutron [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Updated VIF entry in instance network info cache for port 256b055b-e5b6-4f7d-97c7-d21c01586718. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.271471] env[62066]: DEBUG nova.network.neutron [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Updating instance_info_cache with network_info: [{"id": "256b055b-e5b6-4f7d-97c7-d21c01586718", "address": "fa:16:3e:68:17:b4", "network": {"id": "106cfd91-6d32-4e17-8e5d-a147084c0206", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1997972038-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e62ef3f429234c38831e496abac10869", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d0bdd11b-58af-4cc0-9d38-8322e1bb4e74", "external-id": "nsx-vlan-transportzone-398", "segmentation_id": 398, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap256b055b-e5", "ovs_interfaceid": "256b055b-e5b6-4f7d-97c7-d21c01586718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.281044] env[62066]: DEBUG nova.scheduler.client.report [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 137 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1019.281770] env[62066]: DEBUG nova.compute.provider_tree [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 137 to 138 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1019.283031] env[62066]: DEBUG nova.compute.provider_tree [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.345381] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]527e54e9-2620-f1b3-643e-ea09566ba590, 'name': SearchDatastore_Task, 'duration_secs': 0.01541} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.345703] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.345967] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.346232] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.346530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.346596] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.346858] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bf5036f-947a-4f7d-bc11-e246cc14bb8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.358200] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.358425] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1019.359213] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a39ac2a8-e27f-44b6-abf7-b4a4610e6153 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.369774] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1019.369774] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5251a850-7d4a-3cfc-88ee-79698b212ad1" [ 1019.369774] env[62066]: _type = "Task" [ 1019.369774] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.381770] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5251a850-7d4a-3cfc-88ee-79698b212ad1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.640529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.641503] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97206b85-5a8a-4cb1-94c7-ad4452f1733b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.649406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab839f7-8e81-4712-831c-2bded4fb2ef6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.691926] env[62066]: DEBUG nova.network.neutron [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updated VIF entry in instance network info cache for port 567b53ff-9743-4fa0-b6ed-37cab139833f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1019.692597] env[62066]: DEBUG nova.network.neutron [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.774440] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdbb6e4f-8969-46b8-bd43-d6b97cccc6c8 req-ed4fa312-4bd7-41ed-a596-3d98b5da3437 service nova] Releasing lock "refresh_cache-e273b275-2c03-4ac6-919e-46fddde91903" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.786974] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.063s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.787519] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1019.790237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.636s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.793399] env[62066]: INFO nova.compute.claims [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.868526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.868792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.869018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.869592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.869798] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.874771] env[62066]: INFO nova.compute.manager [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Terminating instance [ 1019.876857] env[62066]: DEBUG nova.compute.manager [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.877077] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.877813] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b219617-ca7a-48ee-9b23-1db79a80c7d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.883844] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5251a850-7d4a-3cfc-88ee-79698b212ad1, 'name': SearchDatastore_Task, 'duration_secs': 0.014948} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.884904] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6eae0b5-d849-4eb4-b498-c9cd8d977bdd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.889037] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.889578] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09bf43e5-3fc8-4a66-8f22-4ba9e88aa5b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.892240] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1019.892240] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52903818-2698-0d66-3cc4-77040a8fa073" [ 1019.892240] env[62066]: _type = "Task" [ 1019.892240] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.900862] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52903818-2698-0d66-3cc4-77040a8fa073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.954808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.955070] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.955255] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore2] 08bc5e66-939c-436e-8b7b-dff7f8ff50f4 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.955590] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fddf6ac1-6079-4cf4-8147-7c16e2a61e78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.963082] env[62066]: DEBUG oslo_vmware.api [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1019.963082] env[62066]: value = "task-1156674" [ 1019.963082] env[62066]: _type = "Task" [ 1019.963082] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.971197] env[62066]: DEBUG oslo_vmware.api [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.195769] env[62066]: DEBUG oslo_concurrency.lockutils [req-f37a8815-d672-4e2d-bf2d-56eaf1e5dccb req-af7c8524-f576-4f2d-bf92-c65240c6cf40 service nova] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.251095] env[62066]: DEBUG nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Successfully updated port: 27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.297155] env[62066]: DEBUG nova.compute.utils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1020.301206] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1020.301446] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.342395] env[62066]: DEBUG nova.policy [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0757382f267042ecbc68f0c08607ed93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a5a550637cd748b9ae9988f2cf838c2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.402952] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52903818-2698-0d66-3cc4-77040a8fa073, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.403256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.403516] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] e273b275-2c03-4ac6-919e-46fddde91903/e273b275-2c03-4ac6-919e-46fddde91903.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.403774] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60aaae12-efaf-4454-99f3-62ca9a4901d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.412350] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1020.412350] env[62066]: value = "task-1156675" [ 1020.412350] env[62066]: _type = "Task" [ 1020.412350] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.420520] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.476068] env[62066]: DEBUG oslo_vmware.api [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144297} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.476316] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.476520] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.476707] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.476901] env[62066]: INFO nova.compute.manager [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1020.477177] env[62066]: DEBUG oslo.service.loopingcall [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.477391] env[62066]: DEBUG nova.compute.manager [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.477487] env[62066]: DEBUG nova.network.neutron [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.642512] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Successfully created port: a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.692602] env[62066]: DEBUG nova.compute.manager [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.692718] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.692917] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.693312] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.693561] env[62066]: DEBUG nova.compute.manager [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] No waiting events found dispatching network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.693739] env[62066]: WARNING nova.compute.manager [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received unexpected event network-vif-plugged-27157a0e-629d-45bb-9bdf-b8e235ce8be0 for instance with vm_state active and task_state None. [ 1020.693865] env[62066]: DEBUG nova.compute.manager [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-changed-27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.694038] env[62066]: DEBUG nova.compute.manager [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing instance network info cache due to event network-changed-27157a0e-629d-45bb-9bdf-b8e235ce8be0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.694243] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.694688] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.694688] env[62066]: DEBUG nova.network.neutron [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Refreshing network info cache for port 27157a0e-629d-45bb-9bdf-b8e235ce8be0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.744034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.744289] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb76627c-42ad-45e2-9d7e-95b457d9356c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.753975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.754386] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1020.754386] env[62066]: value = "task-1156676" [ 1020.754386] env[62066]: _type = "Task" [ 1020.754386] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.766678] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.802549] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.927099] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45912} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.927099] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] e273b275-2c03-4ac6-919e-46fddde91903/e273b275-2c03-4ac6-919e-46fddde91903.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.927560] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.927636] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a723bd2-cff2-4a13-a624-912ce05dc986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.935478] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1020.935478] env[62066]: value = "task-1156677" [ 1020.935478] env[62066]: _type = "Task" [ 1020.935478] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.949499] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156677, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.968464] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17351947-440d-4223-8558-7521c1515b1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.976659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08255baa-4f57-4e8d-ac22-ee70cf72dd89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.009032] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fbad67-beb1-4170-9657-f21c7393c767 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.017027] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268d2294-d8a3-4268-b1a2-7be5e043ed0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.031834] env[62066]: DEBUG nova.compute.provider_tree [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.119764] env[62066]: DEBUG nova.compute.manager [req-b54e2d12-5c9e-41e5-a08d-218845dc8215 req-8c90fa8b-b276-4fcd-b10a-49dda67412e4 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Received event network-vif-deleted-f51f939f-f9be-457d-8222-b65f88a1490f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.120024] env[62066]: INFO nova.compute.manager [req-b54e2d12-5c9e-41e5-a08d-218845dc8215 req-8c90fa8b-b276-4fcd-b10a-49dda67412e4 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Neutron deleted interface f51f939f-f9be-457d-8222-b65f88a1490f; detaching it from the instance and deleting it from the info cache [ 1021.120237] env[62066]: DEBUG nova.network.neutron [req-b54e2d12-5c9e-41e5-a08d-218845dc8215 req-8c90fa8b-b276-4fcd-b10a-49dda67412e4 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.265673] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156676, 'name': PowerOffVM_Task, 'duration_secs': 0.207564} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.268206] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.268935] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.269192] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.269381] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.269611] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.269733] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.269902] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.270142] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.270307] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.270474] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.270636] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.270806] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.275981] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae3de1f7-853d-463f-a4f5-02fa76d9519e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.293310] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1021.293310] env[62066]: value = "task-1156678" [ 1021.293310] env[62066]: _type = "Task" [ 1021.293310] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.316685] env[62066]: DEBUG nova.network.neutron [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.446294] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156677, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065096} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.447185] env[62066]: DEBUG nova.network.neutron [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Added VIF to instance network info cache for port 27157a0e-629d-45bb-9bdf-b8e235ce8be0. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 1021.447597] env[62066]: DEBUG nova.network.neutron [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "address": "fa:16:3e:06:2e:5f", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27157a0e-62", "ovs_interfaceid": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.448896] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.449768] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c52aa18-9473-4d21-a284-477b9a1c2361 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.481178] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] e273b275-2c03-4ac6-919e-46fddde91903/e273b275-2c03-4ac6-919e-46fddde91903.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.482067] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c411bb34-e880-44b6-9d91-1297dbb5d9e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.503057] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1021.503057] env[62066]: value = "task-1156679" [ 1021.503057] env[62066]: _type = "Task" [ 1021.503057] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.511301] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.566298] env[62066]: DEBUG nova.scheduler.client.report [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1021.566598] env[62066]: DEBUG nova.compute.provider_tree [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 138 to 139 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1021.566843] env[62066]: DEBUG nova.compute.provider_tree [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.623457] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c1cb35a-5e53-4b48-9c9b-5a3bb5e39171 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.634566] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1370fee5-edae-45d9-8938-0b010a7d3c31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.667174] env[62066]: DEBUG nova.compute.manager [req-b54e2d12-5c9e-41e5-a08d-218845dc8215 req-8c90fa8b-b276-4fcd-b10a-49dda67412e4 service nova] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Detach interface failed, port_id=f51f939f-f9be-457d-8222-b65f88a1490f, reason: Instance 08bc5e66-939c-436e-8b7b-dff7f8ff50f4 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1021.803685] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156678, 'name': ReconfigVM_Task, 'duration_secs': 0.302109} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.804686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4162fb1f-0e69-48d0-99a2-4f4ba959bfc4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.826639] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1021.828858] env[62066]: INFO nova.compute.manager [-] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Took 1.35 seconds to deallocate network for instance. [ 1021.829860] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.829860] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.829993] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.830169] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.830324] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.830476] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.830681] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.831403] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.831403] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.831403] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.831403] env[62066]: DEBUG nova.virt.hardware [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.834124] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69fee0d2-74ea-4b42-8fa4-e65e1d523c84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.842985] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1021.842985] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523e7d51-faca-157c-00a0-7728da64690f" [ 1021.842985] env[62066]: _type = "Task" [ 1021.842985] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.852226] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523e7d51-faca-157c-00a0-7728da64690f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.860127] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.860401] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.860563] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.860761] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.861120] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.861120] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.861335] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.861530] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.861732] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.861917] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.862123] env[62066]: DEBUG nova.virt.hardware [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.863012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ff8230-8f18-4855-833f-9220371502d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.871619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d76657-c2ac-426b-b090-060237922ceb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.950120] env[62066]: DEBUG oslo_concurrency.lockutils [req-1eeda25a-d80d-4fbb-b148-09ebf528bc83 req-edc572da-a402-49fa-9a59-b7359afb3d4d service nova] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.950654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.950836] env[62066]: DEBUG nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1022.014129] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156679, 'name': ReconfigVM_Task, 'duration_secs': 0.304377} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.014563] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Reconfigured VM instance instance-00000068 to attach disk [datastore1] e273b275-2c03-4ac6-919e-46fddde91903/e273b275-2c03-4ac6-919e-46fddde91903.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.015355] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe30cc57-fcf4-4453-abcb-9661bf74e67e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.025280] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1022.025280] env[62066]: value = "task-1156680" [ 1022.025280] env[62066]: _type = "Task" [ 1022.025280] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.034241] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156680, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.072717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.073533] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1022.198217] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Successfully updated port: a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.340412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.340703] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.340937] env[62066]: DEBUG nova.objects.instance [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'resources' on Instance uuid 08bc5e66-939c-436e-8b7b-dff7f8ff50f4 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.356488] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523e7d51-faca-157c-00a0-7728da64690f, 'name': SearchDatastore_Task, 'duration_secs': 0.008412} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.361962] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1022.362492] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26ae8a60-66ca-43fc-80ef-02daf080e09a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.382342] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1022.382342] env[62066]: value = "task-1156681" [ 1022.382342] env[62066]: _type = "Task" [ 1022.382342] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.391081] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.489914] env[62066]: WARNING nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 1022.490183] env[62066]: WARNING nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95 already exists in list: networks containing: ['cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95']. ignoring it [ 1022.490357] env[62066]: WARNING nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] 27157a0e-629d-45bb-9bdf-b8e235ce8be0 already exists in list: port_ids containing: ['27157a0e-629d-45bb-9bdf-b8e235ce8be0']. ignoring it [ 1022.538080] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156680, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.585022] env[62066]: DEBUG nova.compute.utils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1022.585492] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1022.585709] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1022.654171] env[62066]: DEBUG nova.policy [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20873514f0fd4c4496d0cb5b2e500650', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a2bdc9041034d43b33453c202bd6cb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.702215] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.702425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.703263] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1022.896594] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156681, 'name': ReconfigVM_Task, 'duration_secs': 0.280868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.896996] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1022.900317] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc322ba-5649-4ff5-8dba-69000814bd89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.930379] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.933420] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d9b6cff-7749-42c9-ae2c-33bdc66cc64e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.957543] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1022.957543] env[62066]: value = "task-1156682" [ 1022.957543] env[62066]: _type = "Task" [ 1022.957543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.971016] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156682, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.040185] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156680, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.041334] env[62066]: DEBUG nova.network.neutron [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "address": "fa:16:3e:06:2e:5f", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27157a0e-62", "ovs_interfaceid": "27157a0e-629d-45bb-9bdf-b8e235ce8be0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.052078] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6321c96-b561-4baa-941d-eaad6066218d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.062543] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f4b6b8-762a-4841-a8f4-d2b1f4ea22df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.100308] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.104091] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Successfully created port: cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.106903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f4dffb-76af-4eeb-93a6-7eefc9557c1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.116464] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dea2fe7-83ae-461b-a2c5-b457605d5966 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.133236] env[62066]: DEBUG nova.compute.provider_tree [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.240175] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.303719] env[62066]: DEBUG nova.compute.manager [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Received event network-vif-plugged-a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.303922] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.304347] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.304636] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.304837] env[62066]: DEBUG nova.compute.manager [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] No waiting events found dispatching network-vif-plugged-a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1023.305061] env[62066]: WARNING nova.compute.manager [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Received unexpected event network-vif-plugged-a8210a6e-c7ee-4503-b149-d547cf8b33b4 for instance with vm_state building and task_state spawning. [ 1023.305253] env[62066]: DEBUG nova.compute.manager [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Received event network-changed-a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.305416] env[62066]: DEBUG nova.compute.manager [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Refreshing instance network info cache due to event network-changed-a8210a6e-c7ee-4503-b149-d547cf8b33b4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1023.305711] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Acquiring lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.401320] env[62066]: DEBUG nova.network.neutron [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating instance_info_cache with network_info: [{"id": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "address": "fa:16:3e:df:93:9a", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8210a6e-c7", "ovs_interfaceid": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.470515] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.538671] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156680, 'name': Rename_Task, 'duration_secs': 1.146626} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.539089] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.539392] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c937bed-6d40-422a-aae9-be9d29a491d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.543976] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.544697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.544921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.546258] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaa8f05-3d00-4a1a-8841-b9a3a077ddc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.550798] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1023.550798] env[62066]: value = "task-1156683" [ 1023.550798] env[62066]: _type = "Task" [ 1023.550798] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.568494] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.568788] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.568975] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.569205] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.569456] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.569671] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.569899] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.570079] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.570255] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.570417] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.570638] env[62066]: DEBUG nova.virt.hardware [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.579317] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1023.580597] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39a16352-85fb-4b25-8293-f1d34f3c0402 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.599022] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156683, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.606422] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1023.606422] env[62066]: value = "task-1156684" [ 1023.606422] env[62066]: _type = "Task" [ 1023.606422] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.621128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.621397] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.626926] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.637592] env[62066]: DEBUG nova.scheduler.client.report [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.904928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.905282] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Instance network_info: |[{"id": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "address": "fa:16:3e:df:93:9a", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8210a6e-c7", "ovs_interfaceid": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1023.905609] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Acquired lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.905833] env[62066]: DEBUG nova.network.neutron [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Refreshing network info cache for port a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.908051] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:93:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc9714ff-7109-4ea1-9435-b2b3fbdb9e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8210a6e-c7ee-4503-b149-d547cf8b33b4', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.915616] env[62066]: DEBUG oslo.service.loopingcall [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.916582] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1023.916827] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ada6a4e-f7c7-4643-8624-016c9682f5b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.938303] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.938303] env[62066]: value = "task-1156685" [ 1023.938303] env[62066]: _type = "Task" [ 1023.938303] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.947950] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156685, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.968620] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156682, 'name': ReconfigVM_Task, 'duration_secs': 0.783738} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.969039] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to attach disk [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138/ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.969979] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2ef30f-d325-441e-bde1-ec517fcb3661 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.991654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bf0ec8-9021-4847-8260-2b06fbd5fd5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.013136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3301c48d-97a0-4ee1-824f-2524609442d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.034560] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda8a52a-ac5b-4932-a257-3cf1958cea60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.042602] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1024.042893] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dd27c5d-0484-41c7-8741-b8118b8fd260 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.050078] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1024.050078] env[62066]: value = "task-1156686" [ 1024.050078] env[62066]: _type = "Task" [ 1024.050078] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.063455] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.066708] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156683, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.116552] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.117714] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1024.128165] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1024.144096] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.144469] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.144687] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.144948] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.145712] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.145712] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.145712] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.145893] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.146159] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.146351] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.146536] env[62066]: DEBUG nova.virt.hardware [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.147272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.150631] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e436622d-1174-49e3-8a64-e1b416286624 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.160574] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f31f71e-9454-4fd0-8493-2aa498696a5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.180553] env[62066]: INFO nova.scheduler.client.report [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 08bc5e66-939c-436e-8b7b-dff7f8ff50f4 [ 1024.450976] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156685, 'name': CreateVM_Task, 'duration_secs': 0.357643} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.451168] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1024.451824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.451998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.452327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.452656] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5daeda2-b752-4eb7-bd2b-79f7e10ea8d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.458810] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1024.458810] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524cfc3a-0ed8-4438-e25c-e5eff0659d29" [ 1024.458810] env[62066]: _type = "Task" [ 1024.458810] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.467344] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524cfc3a-0ed8-4438-e25c-e5eff0659d29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.564909] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156686, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.570622] env[62066]: DEBUG oslo_vmware.api [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156683, 'name': PowerOnVM_Task, 'duration_secs': 0.825868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.570884] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.571128] env[62066]: INFO nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Took 8.62 seconds to spawn the instance on the hypervisor. [ 1024.571312] env[62066]: DEBUG nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.572069] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005aab74-c12a-4601-8fb4-8688f0807d6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.617508] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.635976] env[62066]: DEBUG nova.network.neutron [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updated VIF entry in instance network info cache for port a8210a6e-c7ee-4503-b149-d547cf8b33b4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1024.636401] env[62066]: DEBUG nova.network.neutron [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating instance_info_cache with network_info: [{"id": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "address": "fa:16:3e:df:93:9a", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8210a6e-c7", "ovs_interfaceid": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.651630] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.651866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.653444] env[62066]: INFO nova.compute.claims [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.687942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59891abe-8fa3-4e36-abd4-a2f3933ac941 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "08bc5e66-939c-436e-8b7b-dff7f8ff50f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.819s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.914520] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Successfully updated port: cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.970449] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524cfc3a-0ed8-4438-e25c-e5eff0659d29, 'name': SearchDatastore_Task, 'duration_secs': 0.01149} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.970685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.970862] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.971110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.971260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.971439] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.971703] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28e992ab-312b-4b18-98b8-001124cc48e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.980931] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.981132] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.981838] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b62977d1-2840-4e26-914e-b6838b553355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.987910] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1024.987910] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5268e38b-426e-d8a7-785b-deb7f8095cda" [ 1024.987910] env[62066]: _type = "Task" [ 1024.987910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.996665] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268e38b-426e-d8a7-785b-deb7f8095cda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.062246] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156686, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.090022] env[62066]: INFO nova.compute.manager [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Took 16.38 seconds to build instance. [ 1025.118603] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.140495] env[62066]: DEBUG oslo_concurrency.lockutils [req-1531da26-b54a-44ae-99af-ac85ef39d6c0 req-52f49458-67df-4f7b-908e-b24d75550d89 service nova] Releasing lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.143174] env[62066]: DEBUG nova.compute.manager [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Received event network-vif-plugged-cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.143421] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.143631] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.143805] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.143976] env[62066]: DEBUG nova.compute.manager [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] No waiting events found dispatching network-vif-plugged-cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.144161] env[62066]: WARNING nova.compute.manager [req-4b073198-9b9a-4935-82c9-95a560228db6 req-b362f06d-c248-4f9e-98b8-c29d36f6b1d9 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Received unexpected event network-vif-plugged-cda4ad76-1cf1-4804-ac63-a7200bd27f67 for instance with vm_state building and task_state spawning. [ 1025.404913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.405222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.417916] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.418145] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.418355] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.499303] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5268e38b-426e-d8a7-785b-deb7f8095cda, 'name': SearchDatastore_Task, 'duration_secs': 0.010229} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.500115] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3409a5d6-8a75-49a9-bb8a-ff218d264c54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.506152] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1025.506152] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]528cc9b6-34e8-5d4b-dca4-fece9571bbc5" [ 1025.506152] env[62066]: _type = "Task" [ 1025.506152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.513766] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528cc9b6-34e8-5d4b-dca4-fece9571bbc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.560220] env[62066]: DEBUG oslo_vmware.api [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156686, 'name': PowerOnVM_Task, 'duration_secs': 1.161927} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.560499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1025.592034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c11c849-1c66-44c1-bfdf-4226729e3fcf tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.892s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.618153] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.737597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "e273b275-2c03-4ac6-919e-46fddde91903" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.737926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.738219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "e273b275-2c03-4ac6-919e-46fddde91903-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.738447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.738633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.740814] env[62066]: INFO nova.compute.manager [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Terminating instance [ 1025.743057] env[62066]: DEBUG nova.compute.manager [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.743057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.743664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092059e4-a4c7-4851-b0ed-b5b262aedb00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.751891] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.752144] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba93f371-6631-4509-b271-4b4bac689aa8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.759417] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1025.759417] env[62066]: value = "task-1156687" [ 1025.759417] env[62066]: _type = "Task" [ 1025.759417] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.770283] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156687, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.793543] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f8aaf5-dd2e-4c71-aa19-6606f12b3873 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.801226] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffb5c67-a3f7-4f14-9b02-9a9a62068c6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.831383] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ddf9c2-41ae-4ed0-b402-f3555b87c400 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.839218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3294fabf-8c23-49ec-a3da-3452b43ffccb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.852404] env[62066]: DEBUG nova.compute.provider_tree [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.907747] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.949258] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1026.019777] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]528cc9b6-34e8-5d4b-dca4-fece9571bbc5, 'name': SearchDatastore_Task, 'duration_secs': 0.010274} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.023321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.023724] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 9c8bf6da-a21e-4566-9c15-4457b9a3de02/9c8bf6da-a21e-4566-9c15-4457b9a3de02.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1026.024098] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5bc684d9-1fcf-47f8-9c7c-ef155b4688a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.032542] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1026.032542] env[62066]: value = "task-1156688" [ 1026.032542] env[62066]: _type = "Task" [ 1026.032542] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.044358] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.120287] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.133247] env[62066]: DEBUG nova.network.neutron [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [{"id": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "address": "fa:16:3e:86:2c:ba", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcda4ad76-1c", "ovs_interfaceid": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.272613] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156687, 'name': PowerOffVM_Task, 'duration_secs': 0.250341} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.272940] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.273112] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.273385] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afe7d29d-22c0-42a0-8447-c06fae4687fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.355361] env[62066]: DEBUG nova.scheduler.client.report [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.387196] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.387446] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.387668] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Deleting the datastore file [datastore1] e273b275-2c03-4ac6-919e-46fddde91903 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.388104] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5912c78b-1da0-45b7-bacf-476401248911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.395910] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for the task: (returnval){ [ 1026.395910] env[62066]: value = "task-1156690" [ 1026.395910] env[62066]: _type = "Task" [ 1026.395910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.406993] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.428607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.544927] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156688, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.614755] env[62066]: INFO nova.compute.manager [None req-7314d1e2-5270-45d4-954e-397e78ebf985 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance to original state: 'active' [ 1026.624219] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.636037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.636235] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Instance network_info: |[{"id": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "address": "fa:16:3e:86:2c:ba", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcda4ad76-1c", "ovs_interfaceid": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.637481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:2c:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bb10726-a946-47b9-b4b5-6916e3f14cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cda4ad76-1cf1-4804-ac63-a7200bd27f67', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.644429] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating folder: Project (1a2bdc9041034d43b33453c202bd6cb4). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1026.644752] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d271e2c9-10d8-48a6-8685-becc39178439 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.656120] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created folder: Project (1a2bdc9041034d43b33453c202bd6cb4) in parent group-v251573. [ 1026.656354] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating folder: Instances. Parent ref: group-v251774. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1026.657302] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0447277a-86d0-4044-81a8-5cc4babf2db2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.667825] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created folder: Instances in parent group-v251774. [ 1026.668471] env[62066]: DEBUG oslo.service.loopingcall [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.668471] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.668597] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5e31bbe-c847-416a-badc-9aeb1229ece4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.688802] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.688802] env[62066]: value = "task-1156693" [ 1026.688802] env[62066]: _type = "Task" [ 1026.688802] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.696792] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156693, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.860562] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.860985] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.864708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.436s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.866345] env[62066]: INFO nova.compute.claims [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.907204] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.045168] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156688, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516006} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.045539] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 9c8bf6da-a21e-4566-9c15-4457b9a3de02/9c8bf6da-a21e-4566-9c15-4457b9a3de02.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.045644] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.045909] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-974e8350-1816-49ab-9a05-4ca5a63123ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.053166] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1027.053166] env[62066]: value = "task-1156694" [ 1027.053166] env[62066]: _type = "Task" [ 1027.053166] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.063357] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156694, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.122842] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.170578] env[62066]: DEBUG nova.compute.manager [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Received event network-changed-cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.171377] env[62066]: DEBUG nova.compute.manager [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Refreshing instance network info cache due to event network-changed-cda4ad76-1cf1-4804-ac63-a7200bd27f67. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1027.171377] env[62066]: DEBUG oslo_concurrency.lockutils [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] Acquiring lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.171377] env[62066]: DEBUG oslo_concurrency.lockutils [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] Acquired lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.171377] env[62066]: DEBUG nova.network.neutron [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Refreshing network info cache for port cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.199912] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156693, 'name': CreateVM_Task, 'duration_secs': 0.447901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.200160] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1027.200917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.201124] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.201447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.201699] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28b3c1d0-56d4-45c4-b46c-62bc74539cec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.207123] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1027.207123] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52bffd42-ce42-49be-3ff2-af46521b91ae" [ 1027.207123] env[62066]: _type = "Task" [ 1027.207123] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.215402] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bffd42-ce42-49be-3ff2-af46521b91ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.372187] env[62066]: DEBUG nova.compute.utils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1027.377104] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1027.377104] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.406929] env[62066]: DEBUG oslo_vmware.api [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Task: {'id': task-1156690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.581189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.407213] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.407405] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.407598] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.407782] env[62066]: INFO nova.compute.manager [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1027.408051] env[62066]: DEBUG oslo.service.loopingcall [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.408256] env[62066]: DEBUG nova.compute.manager [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.408353] env[62066]: DEBUG nova.network.neutron [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1027.536935] env[62066]: DEBUG nova.policy [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91585bdab75c42178c835de912fd3fef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08737e043ede43b58298d1c1a834fa84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.567685] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156694, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082847} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.567685] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.567685] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f2c2bd-dceb-4a02-8637-82894c2d597a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.593285] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 9c8bf6da-a21e-4566-9c15-4457b9a3de02/9c8bf6da-a21e-4566-9c15-4457b9a3de02.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.593856] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-045b6293-08c3-4e0d-89ce-ffe476a6c26c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.618485] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1027.618485] env[62066]: value = "task-1156695" [ 1027.618485] env[62066]: _type = "Task" [ 1027.618485] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.622353] env[62066]: DEBUG oslo_vmware.api [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156684, 'name': ReconfigVM_Task, 'duration_secs': 3.836631} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.626373] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.626966] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1027.641071] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.721100] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52bffd42-ce42-49be-3ff2-af46521b91ae, 'name': SearchDatastore_Task, 'duration_secs': 0.010556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.723673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.723980] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.724165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.724312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.724524] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.724867] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4dc9ede-76cb-4a3b-b638-856e0d57c598 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.734633] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.734852] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.735875] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d323615b-8fed-4c1e-ba50-14f42c168649 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.743921] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1027.743921] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad387d-c32e-e8fb-e5c0-2795a0062035" [ 1027.743921] env[62066]: _type = "Task" [ 1027.743921] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.753878] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad387d-c32e-e8fb-e5c0-2795a0062035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.876827] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.919889] env[62066]: DEBUG nova.network.neutron [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updated VIF entry in instance network info cache for port cda4ad76-1cf1-4804-ac63-a7200bd27f67. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1027.920349] env[62066]: DEBUG nova.network.neutron [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [{"id": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "address": "fa:16:3e:86:2c:ba", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcda4ad76-1c", "ovs_interfaceid": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.957893] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Successfully created port: f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.032350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddca60bd-2222-4ab2-a525-4250c8c5664a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.040998] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18aed594-d41f-4b59-b44d-451a1a8ffbaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.072826] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c0eaf7-002a-4395-964e-ec6505212585 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.081108] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673e87a6-4019-46e4-97a2-08c7124ed8f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.095861] env[62066]: DEBUG nova.compute.provider_tree [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.131827] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156695, 'name': ReconfigVM_Task, 'duration_secs': 0.34615} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.132150] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 9c8bf6da-a21e-4566-9c15-4457b9a3de02/9c8bf6da-a21e-4566-9c15-4457b9a3de02.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.133608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-30b3d7f2-d07d-4877-b9f3-a0a6e671a744 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 10.590s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.134884] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8492dac-6a82-407e-9858-9f94a3ec86f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.147722] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1028.147722] env[62066]: value = "task-1156696" [ 1028.147722] env[62066]: _type = "Task" [ 1028.147722] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.159127] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156696, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.199983] env[62066]: DEBUG nova.network.neutron [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.257097] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad387d-c32e-e8fb-e5c0-2795a0062035, 'name': SearchDatastore_Task, 'duration_secs': 0.010137} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.257999] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28b2f6dd-978a-4c58-abd0-807f6913598f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.265359] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1028.265359] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52635ddd-96cb-3757-a1f9-112c24381256" [ 1028.265359] env[62066]: _type = "Task" [ 1028.265359] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.278819] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52635ddd-96cb-3757-a1f9-112c24381256, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.291166] env[62066]: DEBUG nova.compute.manager [req-2abc12a6-4ddd-4a65-87cb-d8f567f5b233 req-551c9919-5d44-48d2-99de-f2c645e4777b service nova] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Received event network-vif-deleted-256b055b-e5b6-4f7d-97c7-d21c01586718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.385915] env[62066]: INFO nova.virt.block_device [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Booting with volume 6e96e312-38b3-431f-aa8c-cfd4a71a9e9c at /dev/sda [ 1028.423547] env[62066]: DEBUG oslo_concurrency.lockutils [req-590de235-2771-43b2-acd1-8f60971df03d req-cf6d1614-42dc-4b96-90cb-ff2112affc31 service nova] Releasing lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.428268] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3cc97183-5e0a-4d2d-9fa4-7e886e46de5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.438514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0846d7f-f2af-4c18-87d7-068b015a0c88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.474050] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ccd726b-27e6-4ea4-bf45-12885991e39f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.484386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150d8ea5-b144-4f65-a2d6-b39dfdd70b76 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.520700] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20feea6f-96e3-48bf-8f84-21f2609f86fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.528352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fa927e-da56-4748-85f7-bf4db5432122 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.543097] env[62066]: DEBUG nova.virt.block_device [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating existing volume attachment record: 95a098c4-0dd0-45b6-965f-6f2ecbf351b1 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1028.600106] env[62066]: DEBUG nova.scheduler.client.report [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.659475] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156696, 'name': Rename_Task, 'duration_secs': 0.157638} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.659751] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.659994] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cff139f-83d5-4839-a5fa-0bc7442292be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.666527] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1028.666527] env[62066]: value = "task-1156697" [ 1028.666527] env[62066]: _type = "Task" [ 1028.666527] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.674102] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.702494] env[62066]: INFO nova.compute.manager [-] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Took 1.29 seconds to deallocate network for instance. [ 1028.776183] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52635ddd-96cb-3757-a1f9-112c24381256, 'name': SearchDatastore_Task, 'duration_secs': 0.015171} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.776437] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.776719] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a451b1d1-73d3-41ad-b165-23983e48ace5/a451b1d1-73d3-41ad-b165-23983e48ace5.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.777020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8f76b87-fdb2-42e5-b6d3-b09e91c1f04f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.784741] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1028.784741] env[62066]: value = "task-1156698" [ 1028.784741] env[62066]: _type = "Task" [ 1028.784741] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.794310] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.001570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.001570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.001570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.001800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.001973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.004591] env[62066]: INFO nova.compute.manager [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Terminating instance [ 1029.007112] env[62066]: DEBUG nova.compute.manager [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1029.007345] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.007624] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-517a9310-8fb3-4ab1-82fe-685d6543e7df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.016855] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1029.016855] env[62066]: value = "task-1156699" [ 1029.016855] env[62066]: _type = "Task" [ 1029.016855] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.028560] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.105247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.240s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.105765] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.179848] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156697, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.210606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.211071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.211420] env[62066]: DEBUG nova.objects.instance [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lazy-loading 'resources' on Instance uuid e273b275-2c03-4ac6-919e-46fddde91903 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.297878] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500883} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.298145] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] a451b1d1-73d3-41ad-b165-23983e48ace5/a451b1d1-73d3-41ad-b165-23983e48ace5.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.298360] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.298747] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-36143b8c-8f4d-4142-9238-a798027a327f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.307427] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1029.307427] env[62066]: value = "task-1156700" [ 1029.307427] env[62066]: _type = "Task" [ 1029.307427] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.315934] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.531033] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156699, 'name': PowerOffVM_Task, 'duration_secs': 0.322245} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.531641] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.531833] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1029.532158] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251760', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'name': 'volume-9efbca39-3380-44d5-a311-1c9255ac3a78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138', 'attached_at': '2024-10-07T01:00:57.000000', 'detached_at': '', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'serial': '9efbca39-3380-44d5-a311-1c9255ac3a78'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1029.533343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102fba67-c707-4d9a-9437-0d29ff84c7aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.539959] env[62066]: DEBUG nova.compute.manager [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Received event network-vif-plugged-f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.540317] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.540760] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.540944] env[62066]: DEBUG oslo_concurrency.lockutils [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.541255] env[62066]: DEBUG nova.compute.manager [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] No waiting events found dispatching network-vif-plugged-f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.541548] env[62066]: WARNING nova.compute.manager [req-5d678013-a399-42d9-9544-f5d693732369 req-6b660d5a-1e98-4895-bbe3-6efad568258b service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Received unexpected event network-vif-plugged-f557ea42-1811-49ae-90bf-31943c924e96 for instance with vm_state building and task_state block_device_mapping. [ 1029.574254] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb7f92c-d875-456a-b73a-f1e997649277 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.585105] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0732dfb8-2a61-4d6d-92d0-0d552ff2df87 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.620895] env[62066]: DEBUG nova.compute.utils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1029.622711] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.622711] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1029.625759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640ab5db-f527-462b-9ba2-110ec89843bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.628638] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Successfully updated port: f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.645303] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] The volume has not been displaced from its original location: [datastore2] volume-9efbca39-3380-44d5-a311-1c9255ac3a78/volume-9efbca39-3380-44d5-a311-1c9255ac3a78.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1029.650702] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.653126] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d946b1c-137b-4f54-bf91-572ceb1c5991 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.674163] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1029.674163] env[62066]: value = "task-1156701" [ 1029.674163] env[62066]: _type = "Task" [ 1029.674163] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.681350] env[62066]: DEBUG nova.policy [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12b0fb4ac6a4a0ca7b662f5f2ddab6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c1a91ea6e0b4b2da6a16f327bc77a26', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.683491] env[62066]: DEBUG oslo_vmware.api [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156697, 'name': PowerOnVM_Task, 'duration_secs': 0.58795} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.683724] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1029.683945] env[62066]: INFO nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Took 7.86 seconds to spawn the instance on the hypervisor. [ 1029.684153] env[62066]: DEBUG nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.685256] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3da734-08ea-4f09-90e8-94a0ddabb9da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.690668] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156701, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.818651] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.2166} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.818955] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.819815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2042770a-cc43-4150-8109-87a2f6b3a61b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.840172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.840473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.851073] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] a451b1d1-73d3-41ad-b165-23983e48ace5/a451b1d1-73d3-41ad-b165-23983e48ace5.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.855337] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1d00a4e-21a0-4248-99b0-e1c5908bf1b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.876954] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1029.876954] env[62066]: value = "task-1156702" [ 1029.876954] env[62066]: _type = "Task" [ 1029.876954] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.888712] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156702, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.907802] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fdf2e9-50c2-41c8-b0e7-17b2f7757279 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.916613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0acd53-7f26-4917-94c1-a8a9e87d7397 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.958187] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Successfully created port: 65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.958187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1d6dbe-e054-4d63-9b2d-37828eee2931 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.967393] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6893bb-5216-44c9-911f-7062902d87ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.984183] env[62066]: DEBUG nova.compute.provider_tree [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.123503] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.134610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.134783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.134939] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.184521] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156701, 'name': ReconfigVM_Task, 'duration_secs': 0.247117} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.184828] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1030.189925] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb216152-540c-4d87-b530-c487d49b899c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.211023] env[62066]: INFO nova.compute.manager [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Took 19.42 seconds to build instance. [ 1030.214048] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1030.214048] env[62066]: value = "task-1156703" [ 1030.214048] env[62066]: _type = "Task" [ 1030.214048] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.223423] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156703, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.352947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.353193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.354149] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078fb381-f116-40c0-9e42-83f1d7c4fd23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.374752] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d629b79-ef57-4bd7-b3c4-ec9fd79ebf77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.386994] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156702, 'name': ReconfigVM_Task, 'duration_secs': 0.417396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.404089] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfigured VM instance instance-0000006a to attach disk [datastore2] a451b1d1-73d3-41ad-b165-23983e48ace5/a451b1d1-73d3-41ad-b165-23983e48ace5.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.410438] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1030.410745] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad8f9be3-164d-4419-bc6e-a1f1ac13fe4f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.412591] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d5a7874-4bb1-4342-810a-69559c523b70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.432198] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1030.432198] env[62066]: value = "task-1156704" [ 1030.432198] env[62066]: _type = "Task" [ 1030.432198] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.433646] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1030.433646] env[62066]: value = "task-1156705" [ 1030.433646] env[62066]: _type = "Task" [ 1030.433646] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.449180] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.452429] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156704, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.487871] env[62066]: DEBUG nova.scheduler.client.report [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.667227] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.671338] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.671848] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.672081] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.672247] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.672435] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.672729] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.672979] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.673229] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.673403] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.673585] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.673754] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.673950] env[62066]: DEBUG nova.virt.hardware [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.675152] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf4d09e-79f0-4da4-bdd5-8729cf9365c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.688048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64458db-c232-4d07-9970-447b00c24d76 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.715790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5459459a-7c73-4d52-b5e7-97b08edfe78d tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.932s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.724796] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156703, 'name': ReconfigVM_Task, 'duration_secs': 0.225184} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.725130] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251760', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'name': 'volume-9efbca39-3380-44d5-a311-1c9255ac3a78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138', 'attached_at': '2024-10-07T01:00:57.000000', 'detached_at': '', 'volume_id': '9efbca39-3380-44d5-a311-1c9255ac3a78', 'serial': '9efbca39-3380-44d5-a311-1c9255ac3a78'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1030.725430] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.726315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9786ac9-c833-464c-8dea-f01ac60020ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.734260] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.734260] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1dd19a7f-27d5-478f-a442-3b8a4f68ab99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.802012] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.802012] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.802012] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleting the datastore file [datastore1] ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.802296] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e851c1cc-17b4-46ac-be10-26a7dddd1c3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.810318] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1030.810318] env[62066]: value = "task-1156707" [ 1030.810318] env[62066]: _type = "Task" [ 1030.810318] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.818728] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.885653] env[62066]: DEBUG nova.network.neutron [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.944435] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156704, 'name': Rename_Task, 'duration_secs': 0.176008} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.945225] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1030.945457] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04b010bb-aeb9-4b93-ab5e-7f4e3be1d285 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.950647] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.957044] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1030.957044] env[62066]: value = "task-1156708" [ 1030.957044] env[62066]: _type = "Task" [ 1030.957044] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.965303] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.992855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.014218] env[62066]: INFO nova.scheduler.client.report [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Deleted allocations for instance e273b275-2c03-4ac6-919e-46fddde91903 [ 1031.133695] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.161072] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.161175] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.161367] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.161573] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.161729] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.161883] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.162109] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.162276] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.162448] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.162614] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.162790] env[62066]: DEBUG nova.virt.hardware [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.163680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbe317b-d2bb-4aa7-a259-cef1454c17c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.172912] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f70b874-1df6-4e67-8d78-59388931826f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.321527] env[62066]: DEBUG oslo_vmware.api [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.321594] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.321818] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.322088] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.322328] env[62066]: INFO nova.compute.manager [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Took 2.31 seconds to destroy the instance on the hypervisor. [ 1031.322639] env[62066]: DEBUG oslo.service.loopingcall [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.322864] env[62066]: DEBUG nova.compute.manager [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.322977] env[62066]: DEBUG nova.network.neutron [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1031.375156] env[62066]: DEBUG nova.compute.manager [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Received event network-vif-plugged-65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.375269] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.375454] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.375627] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.375897] env[62066]: DEBUG nova.compute.manager [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] No waiting events found dispatching network-vif-plugged-65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.376284] env[62066]: WARNING nova.compute.manager [req-8f5e0e6f-7a92-4cfa-b644-7536a8bf2d3c req-584519b8-4ce7-497d-bd34-8f3829f8e366 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Received unexpected event network-vif-plugged-65045fb9-30a1-493e-a710-cb21606de596 for instance with vm_state building and task_state spawning. [ 1031.388181] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.388588] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Instance network_info: |[{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.389158] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:e9:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f557ea42-1811-49ae-90bf-31943c924e96', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.397880] env[62066]: DEBUG oslo.service.loopingcall [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.398644] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.398901] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96525dfd-ae5f-4dbb-b036-751ad609b42a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.420952] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.420952] env[62066]: value = "task-1156709" [ 1031.420952] env[62066]: _type = "Task" [ 1031.420952] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.430793] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156709, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.448653] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.469901] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156708, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.522972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de4deeeb-eed0-461f-bcb3-11b4c6e50f42 tempest-InstanceActionsV221TestJSON-1902553891 tempest-InstanceActionsV221TestJSON-1902553891-project-member] Lock "e273b275-2c03-4ac6-919e-46fddde91903" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.785s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.573591] env[62066]: DEBUG nova.compute.manager [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Received event network-changed-f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.573796] env[62066]: DEBUG nova.compute.manager [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Refreshing instance network info cache due to event network-changed-f557ea42-1811-49ae-90bf-31943c924e96. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.573929] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.574113] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.574298] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Refreshing network info cache for port f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.932962] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156709, 'name': CreateVM_Task, 'duration_secs': 0.331489} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.933932] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Successfully updated port: 65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.935187] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.935947] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'device_type': None, 'boot_index': 0, 'attachment_id': '95a098c4-0dd0-45b6-965f-6f2ecbf351b1', 'mount_device': '/dev/sda', 'guest_format': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251769', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'name': 'volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762', 'attached_at': '', 'detached_at': '', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'serial': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c'}, 'volume_type': None}], 'swap': None} {{(pid=62066) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1031.936217] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Root volume attach. Driver type: vmdk {{(pid=62066) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1031.937282] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92736b00-d88b-4063-bb7b-47313973b60e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.953916] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06af60b8-0c1e-4fde-9795-983ca98d061f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.956696] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.959251] env[62066]: DEBUG nova.compute.manager [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Received event network-changed-65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.959452] env[62066]: DEBUG nova.compute.manager [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Refreshing instance network info cache due to event network-changed-65045fb9-30a1-493e-a710-cb21606de596. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.959717] env[62066]: DEBUG oslo_concurrency.lockutils [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] Acquiring lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.959893] env[62066]: DEBUG oslo_concurrency.lockutils [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] Acquired lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.960106] env[62066]: DEBUG nova.network.neutron [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Refreshing network info cache for port 65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.968454] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7614162d-36c3-49bf-9e51-da063d09dff4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.975327] env[62066]: DEBUG oslo_vmware.api [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156708, 'name': PowerOnVM_Task, 'duration_secs': 0.556361} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.975962] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.976192] env[62066]: INFO nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Took 7.86 seconds to spawn the instance on the hypervisor. [ 1031.976377] env[62066]: DEBUG nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.977197] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2b18d6-b9d7-40b0-b8e7-254997f490f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.982928] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-52c847e2-1904-4e33-90fc-b161b2883a96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.994861] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1031.994861] env[62066]: value = "task-1156710" [ 1031.994861] env[62066]: _type = "Task" [ 1031.994861] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.005364] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156710, 'name': RelocateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.282558] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updated VIF entry in instance network info cache for port f557ea42-1811-49ae-90bf-31943c924e96. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.283143] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.354810] env[62066]: DEBUG nova.network.neutron [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.437643] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.452705] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.506229] env[62066]: INFO nova.compute.manager [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Took 20.36 seconds to build instance. [ 1032.512116] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156710, 'name': RelocateVM_Task, 'duration_secs': 0.377783} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.513336] env[62066]: DEBUG nova.network.neutron [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1032.515336] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1032.515708] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251769', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'name': 'volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762', 'attached_at': '', 'detached_at': '', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'serial': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1032.516844] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc96d220-25dc-4ed7-bac1-1a88aa0499f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.537390] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8434b29-5fbe-420b-81e5-002da0598486 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.561155] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c/volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.563872] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-359c52cc-cf00-4a18-bf4c-d81e8b36553d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.584815] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1032.584815] env[62066]: value = "task-1156711" [ 1032.584815] env[62066]: _type = "Task" [ 1032.584815] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.593960] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.614942] env[62066]: DEBUG nova.network.neutron [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.786347] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.786644] env[62066]: DEBUG nova.compute.manager [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Received event network-changed-a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.786822] env[62066]: DEBUG nova.compute.manager [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Refreshing instance network info cache due to event network-changed-a8210a6e-c7ee-4503-b149-d547cf8b33b4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.787068] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Acquiring lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.787223] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Acquired lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.787389] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Refreshing network info cache for port a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.857674] env[62066]: INFO nova.compute.manager [-] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Took 1.53 seconds to deallocate network for instance. [ 1032.959422] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.009156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4b68b1fb-18d8-4d07-9ff8-ea8d8d32bd68 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.874s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.096187] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156711, 'name': ReconfigVM_Task, 'duration_secs': 0.260079} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.096477] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c/volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.102726] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-371d82f0-5f8c-4c8f-903a-0d1dd6acb933 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.119232] env[62066]: DEBUG oslo_concurrency.lockutils [req-0413fa05-71b7-4d17-a493-aa6b118998d4 req-cf044e05-a5f6-4f71-8bbe-9bcf0bdfffa3 service nova] Releasing lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.120806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.120973] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.122294] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1033.122294] env[62066]: value = "task-1156712" [ 1033.122294] env[62066]: _type = "Task" [ 1033.122294] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.136270] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156712, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.420763] env[62066]: INFO nova.compute.manager [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Took 0.56 seconds to detach 1 volumes for instance. [ 1033.453851] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.616829] env[62066]: DEBUG nova.compute.manager [req-530c3ab7-b7a2-4fe6-8d6d-169b6a86d6d2 req-37e420e9-b340-4af0-be0e-e19abf83d954 service nova] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Received event network-vif-deleted-f3c50a8a-9b6d-45c3-a6f7-f745d70b8aef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.636239] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156712, 'name': ReconfigVM_Task, 'duration_secs': 0.14823} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.637507] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251769', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'name': 'volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762', 'attached_at': '', 'detached_at': '', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'serial': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1033.638683] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb0bbeaf-2f82-48d6-b782-62121b068660 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.648498] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1033.648498] env[62066]: value = "task-1156713" [ 1033.648498] env[62066]: _type = "Task" [ 1033.648498] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.663576] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156713, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.668201] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.718251] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updated VIF entry in instance network info cache for port a8210a6e-c7ee-4503-b149-d547cf8b33b4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1033.719189] env[62066]: DEBUG nova.network.neutron [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating instance_info_cache with network_info: [{"id": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "address": "fa:16:3e:df:93:9a", "network": {"id": "1b2c8c6e-6c6a-4845-8cc4-aeac415b7a32", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1537257567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5a550637cd748b9ae9988f2cf838c2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc9714ff-7109-4ea1-9435-b2b3fbdb9e81", "external-id": "nsx-vlan-transportzone-887", "segmentation_id": 887, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8210a6e-c7", "ovs_interfaceid": "a8210a6e-c7ee-4503-b149-d547cf8b33b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.829591] env[62066]: DEBUG nova.network.neutron [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [{"id": "65045fb9-30a1-493e-a710-cb21606de596", "address": "fa:16:3e:44:6a:ac", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65045fb9-30", "ovs_interfaceid": "65045fb9-30a1-493e-a710-cb21606de596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.932587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.932587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.932905] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.955188] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.964026] env[62066]: INFO nova.scheduler.client.report [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted allocations for instance ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138 [ 1033.985638] env[62066]: DEBUG nova.compute.manager [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Received event network-changed-cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.985928] env[62066]: DEBUG nova.compute.manager [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Refreshing instance network info cache due to event network-changed-cda4ad76-1cf1-4804-ac63-a7200bd27f67. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1033.986390] env[62066]: DEBUG oslo_concurrency.lockutils [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] Acquiring lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.986626] env[62066]: DEBUG oslo_concurrency.lockutils [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] Acquired lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.986833] env[62066]: DEBUG nova.network.neutron [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Refreshing network info cache for port cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.166987] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156713, 'name': Rename_Task, 'duration_secs': 0.149791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.169783] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.169783] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf185d06-c117-4de8-bb5c-9743b8098b42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.177380] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1034.177380] env[62066]: value = "task-1156714" [ 1034.177380] env[62066]: _type = "Task" [ 1034.177380] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.189832] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.224967] env[62066]: DEBUG oslo_concurrency.lockutils [req-45b18da9-3e96-4d1d-8220-e7cfb9f0562d req-318532ea-056e-4a99-a8e0-868f8638d8ca service nova] Releasing lock "refresh_cache-9c8bf6da-a21e-4566-9c15-4457b9a3de02" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.332429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.333192] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance network_info: |[{"id": "65045fb9-30a1-493e-a710-cb21606de596", "address": "fa:16:3e:44:6a:ac", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65045fb9-30", "ovs_interfaceid": "65045fb9-30a1-493e-a710-cb21606de596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1034.334193] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:6a:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65045fb9-30a1-493e-a710-cb21606de596', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.342756] env[62066]: DEBUG oslo.service.loopingcall [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.343182] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1034.343543] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0dec168-c12e-4f93-9d15-f5ff892a530e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.365215] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.365215] env[62066]: value = "task-1156715" [ 1034.365215] env[62066]: _type = "Task" [ 1034.365215] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.374347] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156715, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.457619] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.474382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cfefaf5b-833a-4037-955e-3b08e739e35b tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.473s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.689902] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156714, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.801357] env[62066]: DEBUG nova.network.neutron [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updated VIF entry in instance network info cache for port cda4ad76-1cf1-4804-ac63-a7200bd27f67. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.801694] env[62066]: DEBUG nova.network.neutron [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [{"id": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "address": "fa:16:3e:86:2c:ba", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcda4ad76-1c", "ovs_interfaceid": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.877697] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156715, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.955527] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.189376] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156714, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.304609] env[62066]: DEBUG oslo_concurrency.lockutils [req-9ad057fc-a4d4-4c6e-bbd6-e247fe541696 req-176d2801-0cf1-4c21-aaf6-51f090329bb1 service nova] Releasing lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.344170] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.344430] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.378499] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156715, 'name': CreateVM_Task, 'duration_secs': 0.590838} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.378499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1035.379256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.379431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.379774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.380058] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f438102-bdf2-4dc4-97eb-153adee60bf9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.385657] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1035.385657] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52201272-f4dd-0935-0b85-b236021d120e" [ 1035.385657] env[62066]: _type = "Task" [ 1035.385657] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.394807] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52201272-f4dd-0935-0b85-b236021d120e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.457404] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.688757] env[62066]: DEBUG oslo_vmware.api [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156714, 'name': PowerOnVM_Task, 'duration_secs': 1.01438} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.689121] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.689349] env[62066]: INFO nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Took 5.02 seconds to spawn the instance on the hypervisor. [ 1035.689536] env[62066]: DEBUG nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.690371] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2284359e-61d8-4524-9b70-7ad72ae74578 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.846668] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.896709] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52201272-f4dd-0935-0b85-b236021d120e, 'name': SearchDatastore_Task, 'duration_secs': 0.016382} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.897104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.897467] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.897979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.897979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.898138] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.898368] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8954c3a-947e-46cc-b142-df67246ced88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.908234] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.908423] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.909144] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81e11332-6aae-4a89-a748-57538d26ee71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.914407] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1035.914407] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c705c3-2143-e470-10af-9ff80bfc22f1" [ 1035.914407] env[62066]: _type = "Task" [ 1035.914407] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.924102] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c705c3-2143-e470-10af-9ff80bfc22f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.954467] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.208025] env[62066]: INFO nova.compute.manager [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Took 11.57 seconds to build instance. [ 1036.373133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.373428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.375035] env[62066]: INFO nova.compute.claims [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.426824] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c705c3-2143-e470-10af-9ff80bfc22f1, 'name': SearchDatastore_Task, 'duration_secs': 0.032349} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.427720] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9074fc48-3c31-4850-a913-6ce9f2d04eea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.433426] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1036.433426] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cc1e-312d-2003-01d8-c3035b9423f8" [ 1036.433426] env[62066]: _type = "Task" [ 1036.433426] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.441904] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cc1e-312d-2003-01d8-c3035b9423f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.454491] env[62066]: DEBUG oslo_vmware.api [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156705, 'name': ReconfigVM_Task, 'duration_secs': 5.900655} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.454795] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.455029] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1036.709244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e75aeda3-92f8-4de5-82bd-8a058e53802b tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.088s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.944439] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5265cc1e-312d-2003-01d8-c3035b9423f8, 'name': SearchDatastore_Task, 'duration_secs': 0.013247} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.944721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.945033] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.945317] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02aa7d34-4f20-4e65-a4d2-c14ed9863e71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.953647] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1036.953647] env[62066]: value = "task-1156716" [ 1036.953647] env[62066]: _type = "Task" [ 1036.953647] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.964453] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.330206] env[62066]: DEBUG nova.compute.manager [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.331869] env[62066]: DEBUG nova.compute.manager [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing instance network info cache due to event network-changed-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.331869] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] Acquiring lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.331869] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] Acquired lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.331869] env[62066]: DEBUG nova.network.neutron [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Refreshing network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1037.468226] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156716, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.520295] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c169652-4661-4048-9cbf-fbd77be140b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.528257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8272e76a-3183-4386-bce8-95dc802d65f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.560059] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31408bd7-e9f3-4b6b-ab67-131fc5fca1df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.568725] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedd7f35-0d58-4083-91d2-0a2881cca697 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.585483] env[62066]: DEBUG nova.compute.provider_tree [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.837276] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.837543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquired lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.837659] env[62066]: DEBUG nova.network.neutron [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.966920] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156716, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.967258] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.967505] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.968093] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7d7d8c2-325d-40ed-806d-5f61ca965ef4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.975819] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1037.975819] env[62066]: value = "task-1156717" [ 1037.975819] env[62066]: _type = "Task" [ 1037.975819] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.985302] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156717, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.089886] env[62066]: DEBUG nova.scheduler.client.report [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.130371] env[62066]: DEBUG nova.network.neutron [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updated VIF entry in instance network info cache for port 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1038.130736] env[62066]: DEBUG nova.network.neutron [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [{"id": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "address": "fa:16:3e:ed:9c:41", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05f9bc18-80", "ovs_interfaceid": "05f9bc18-80b0-42b9-ba2b-8e552a0a77ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.273086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.273466] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.273738] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.274021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.274246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.276772] env[62066]: INFO nova.compute.manager [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Terminating instance [ 1038.279856] env[62066]: DEBUG nova.compute.manager [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.280072] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.280954] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38cace91-4c94-4c98-8d35-7eadda39a2ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.292108] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.292397] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f7c5559-b1db-43d0-80ad-6b60d689b0d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.300633] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1038.300633] env[62066]: value = "task-1156718" [ 1038.300633] env[62066]: _type = "Task" [ 1038.300633] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.310062] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.486698] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156717, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066102} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.486967] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.487763] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a488cb-5489-4c44-91d0-c1ee1882e0df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.511642] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.512357] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5af73738-527d-4051-b8b4-282eb9f1d96c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.534492] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1038.534492] env[62066]: value = "task-1156719" [ 1038.534492] env[62066]: _type = "Task" [ 1038.534492] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.545356] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.573911] env[62066]: INFO nova.network.neutron [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Port 27157a0e-629d-45bb-9bdf-b8e235ce8be0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1038.574396] env[62066]: DEBUG nova.network.neutron [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [{"id": "567b53ff-9743-4fa0-b6ed-37cab139833f", "address": "fa:16:3e:86:8c:d9", "network": {"id": "cf4bfe8a-07ed-41b0-b0f9-c2ced78a2e95", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-496119854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42219a58a1514265b9d0b515eb517933", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "65497291-07f3-434c-bd42-657a0cb03365", "external-id": "nsx-vlan-transportzone-279", "segmentation_id": 279, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap567b53ff-97", "ovs_interfaceid": "567b53ff-9743-4fa0-b6ed-37cab139833f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.596897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.597556] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.633514] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bef1f10-40e8-48d9-92a2-64b5757647f3 req-213dfa08-a6a7-4daa-bcea-2d4132bd0471 service nova] Releasing lock "refresh_cache-5db50c22-048b-4cce-962a-3df1262f6e4f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.813086] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156718, 'name': PowerOffVM_Task, 'duration_secs': 0.485723} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.813086] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.813086] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.813309] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dfefff7-74f4-4680-a242-0a7456d46dba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.870939] env[62066]: DEBUG nova.compute.manager [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1038.922655] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.922918] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.923148] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleting the datastore file [datastore2] 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.923418] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8cdecb74-b338-4329-aa1d-84ac68ad2a65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.930969] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1038.930969] env[62066]: value = "task-1156721" [ 1038.930969] env[62066]: _type = "Task" [ 1038.930969] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.939603] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.048607] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.077695] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Releasing lock "refresh_cache-045c2efa-e2fe-4d51-b267-e2b380fcf1d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.102423] env[62066]: DEBUG nova.compute.utils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1039.103860] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.104058] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1039.161899] env[62066]: DEBUG nova.policy [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c41065776f39438a8f6d427c0f62b93f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c468d5ba348d437f97a74e0da70bb42e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.358835] env[62066]: DEBUG nova.compute.manager [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Received event network-changed-f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.359033] env[62066]: DEBUG nova.compute.manager [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Refreshing instance network info cache due to event network-changed-f557ea42-1811-49ae-90bf-31943c924e96. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.359230] env[62066]: DEBUG oslo_concurrency.lockutils [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.359412] env[62066]: DEBUG oslo_concurrency.lockutils [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.359582] env[62066]: DEBUG nova.network.neutron [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Refreshing network info cache for port f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.392029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.392312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.442106] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.496538] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Successfully created port: 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.546153] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.582433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36130538-744e-42ff-aa1b-2d3f9cc43f50 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "interface-045c2efa-e2fe-4d51-b267-e2b380fcf1d2-27157a0e-629d-45bb-9bdf-b8e235ce8be0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.742s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.607576] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.896874] env[62066]: INFO nova.compute.claims [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.942967] env[62066]: DEBUG oslo_vmware.api [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.559431} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.942967] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.943175] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.943212] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.943384] env[62066]: INFO nova.compute.manager [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1039.943627] env[62066]: DEBUG oslo.service.loopingcall [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.943816] env[62066]: DEBUG nova.compute.manager [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.943908] env[62066]: DEBUG nova.network.neutron [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1040.047704] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.150890] env[62066]: DEBUG nova.network.neutron [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updated VIF entry in instance network info cache for port f557ea42-1811-49ae-90bf-31943c924e96. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.151373] env[62066]: DEBUG nova.network.neutron [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.403480] env[62066]: INFO nova.compute.resource_tracker [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating resource usage from migration 51f468c3-4a2a-4004-97ba-528e090fd6d3 [ 1040.547414] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.548703] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641c9fed-fc91-4f69-ab9a-d5294110b887 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.557994] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4ff085-5d05-4cd6-a0f8-22bd04a8d0db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.592583] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2732abff-2f91-4f20-b12f-8868d5e661ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.601290] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a29725f-bc6a-4545-843d-7f6c56d85d9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.615853] env[62066]: DEBUG nova.compute.provider_tree [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.619084] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.643930] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.644216] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.644382] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.644569] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.644722] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.644930] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.645181] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.645350] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.645524] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.645694] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.645878] env[62066]: DEBUG nova.virt.hardware [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.646766] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50ebd2d-c183-4e63-a809-b0c123a7067a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.655991] env[62066]: DEBUG oslo_concurrency.lockutils [req-41378d01-a640-470e-88b2-3be8b655c2ee req-557ed9e0-f072-4daa-8a4d-bda7548b20aa service nova] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.657742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1297b6e-e95e-4515-a991-fb1a8fdb076d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.887652] env[62066]: DEBUG nova.compute.manager [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.887883] env[62066]: DEBUG oslo_concurrency.lockutils [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.888113] env[62066]: DEBUG oslo_concurrency.lockutils [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.888291] env[62066]: DEBUG oslo_concurrency.lockutils [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.888461] env[62066]: DEBUG nova.compute.manager [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] No waiting events found dispatching network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1040.888628] env[62066]: WARNING nova.compute.manager [req-f842b5d7-3dd7-4e0e-ba03-699f6e779773 req-b7e32e98-9169-4fef-bcc8-907dbf37a767 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received unexpected event network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f for instance with vm_state building and task_state spawning. [ 1040.979945] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Successfully updated port: 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.047889] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156719, 'name': ReconfigVM_Task, 'duration_secs': 2.13763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.048302] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.048971] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c1d15fa-4704-45d9-b857-e755969b4fee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.055816] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1041.055816] env[62066]: value = "task-1156722" [ 1041.055816] env[62066]: _type = "Task" [ 1041.055816] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.063926] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156722, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.123210] env[62066]: DEBUG nova.scheduler.client.report [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.208489] env[62066]: DEBUG nova.network.neutron [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.382694] env[62066]: DEBUG nova.compute.manager [req-f72d0c5d-5139-4d5e-9a65-0fd86cacf2a8 req-60720792-7129-44cf-93eb-05e601cdc537 service nova] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Received event network-vif-deleted-567b53ff-9743-4fa0-b6ed-37cab139833f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.484523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.484749] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.485009] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1041.566207] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156722, 'name': Rename_Task, 'duration_secs': 0.161579} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.566494] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.566746] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-012667ff-4426-41b2-9169-5020d5ffb908 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.573882] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1041.573882] env[62066]: value = "task-1156723" [ 1041.573882] env[62066]: _type = "Task" [ 1041.573882] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.583536] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156723, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.628328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.628551] env[62066]: INFO nova.compute.manager [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Migrating [ 1041.711235] env[62066]: INFO nova.compute.manager [-] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Took 1.77 seconds to deallocate network for instance. [ 1042.018533] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1042.084195] env[62066]: DEBUG oslo_vmware.api [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156723, 'name': PowerOnVM_Task, 'duration_secs': 0.499675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.084469] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.084680] env[62066]: INFO nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Took 10.95 seconds to spawn the instance on the hypervisor. [ 1042.084884] env[62066]: DEBUG nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.085869] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f155cb1a-16e3-4e06-878e-59e3f62213e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.143322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.143553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.143734] env[62066]: DEBUG nova.network.neutron [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.200027] env[62066]: DEBUG nova.network.neutron [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.218740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.219015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.219242] env[62066]: DEBUG nova.objects.instance [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'resources' on Instance uuid 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.601964] env[62066]: INFO nova.compute.manager [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Took 16.19 seconds to build instance. [ 1042.702420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.702763] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance network_info: |[{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.703240] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:b2:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '50c2cd9e-175e-4926-86cb-f67b3875281f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.710734] env[62066]: DEBUG oslo.service.loopingcall [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.713131] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.713376] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aeddf775-7c51-4ea1-ac4f-ffa9425942bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.740266] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.740266] env[62066]: value = "task-1156724" [ 1042.740266] env[62066]: _type = "Task" [ 1042.740266] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.750325] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156724, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.889890] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4c0793-6b80-4362-b9dd-3d62510760f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.898555] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8de1c7-772e-4e56-afc0-43513fcdf9ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.933928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074bd111-7627-4d6d-aaf9-6e4a2ee88521 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.937471] env[62066]: DEBUG nova.compute.manager [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.937660] env[62066]: DEBUG nova.compute.manager [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing instance network info cache due to event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.937880] env[62066]: DEBUG oslo_concurrency.lockutils [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.938037] env[62066]: DEBUG oslo_concurrency.lockutils [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.938208] env[62066]: DEBUG nova.network.neutron [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.939977] env[62066]: DEBUG nova.network.neutron [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.948723] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728770c3-060d-4516-965c-b0e4bbb1694b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.963938] env[62066]: DEBUG nova.compute.provider_tree [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.104584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f35d81e8-477f-4a8a-a14b-1fbc62a30043 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.699s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.250676] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156724, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.445189] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.467565] env[62066]: DEBUG nova.scheduler.client.report [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.685527] env[62066]: DEBUG nova.network.neutron [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updated VIF entry in instance network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1043.685921] env[62066]: DEBUG nova.network.neutron [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.753639] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156724, 'name': CreateVM_Task, 'duration_secs': 0.534794} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.753910] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.754908] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.755228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.755743] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.756137] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9158bc2b-994e-4fc0-91d3-2c5b2a78a58c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.762621] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1043.762621] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]524cdf30-5540-0e53-baee-1b0bee2991c2" [ 1043.762621] env[62066]: _type = "Task" [ 1043.762621] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.774557] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524cdf30-5540-0e53-baee-1b0bee2991c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.975821] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.982065] env[62066]: DEBUG nova.compute.manager [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1043.996995] env[62066]: INFO nova.scheduler.client.report [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted allocations for instance 045c2efa-e2fe-4d51-b267-e2b380fcf1d2 [ 1044.188959] env[62066]: DEBUG oslo_concurrency.lockutils [req-503822a7-abfd-4460-bac8-d9c085b27dcb req-d8211864-5333-4c36-b518-1024249cb394 service nova] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.274613] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]524cdf30-5540-0e53-baee-1b0bee2991c2, 'name': SearchDatastore_Task, 'duration_secs': 0.011256} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.275010] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.275274] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.275517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.275667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.275847] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.276159] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d1f3adf-49dd-4d9a-80fc-7e5656a50ed0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.285386] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.285577] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.286332] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fccc3943-f7c8-48e4-bb17-50855ab95e0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.291751] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1044.291751] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529a37fa-86d4-f5c8-ab12-b1a013939a3f" [ 1044.291751] env[62066]: _type = "Task" [ 1044.291751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.301262] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529a37fa-86d4-f5c8-ab12-b1a013939a3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.499311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.499645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.503454] env[62066]: DEBUG oslo_concurrency.lockutils [None req-013bb680-a82e-490d-8d06-346c5362117d tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "045c2efa-e2fe-4d51-b267-e2b380fcf1d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.230s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.803150] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529a37fa-86d4-f5c8-ab12-b1a013939a3f, 'name': SearchDatastore_Task, 'duration_secs': 0.010483} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.803968] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a036eca-b464-4a8c-b218-d8178665b4f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.809745] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1044.809745] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52fef0f2-4a21-40a0-5eb3-6a8a24629d47" [ 1044.809745] env[62066]: _type = "Task" [ 1044.809745] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.817986] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fef0f2-4a21-40a0-5eb3-6a8a24629d47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.961461] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de82a2ba-9bd9-440f-953d-8f55f6bdf979 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.981401] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.006675] env[62066]: INFO nova.compute.claims [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1045.012917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.013211] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.013421] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "6e0065b0-da4f-4288-a858-8648572f8148-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.013612] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.013782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.015771] env[62066]: INFO nova.compute.manager [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Terminating instance [ 1045.017523] env[62066]: DEBUG nova.compute.manager [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.017718] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.018552] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f605322-9740-4da4-b6bf-8ac2b0e74c9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.026685] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.027609] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5302f52-1a82-4131-b93e-79d5ff8a0ae0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.035658] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1045.035658] env[62066]: value = "task-1156725" [ 1045.035658] env[62066]: _type = "Task" [ 1045.035658] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.045167] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156725, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.320806] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fef0f2-4a21-40a0-5eb3-6a8a24629d47, 'name': SearchDatastore_Task, 'duration_secs': 0.010331} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.321110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.321377] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.321646] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1f590a9-7fa7-4907-a738-1eb1ca45ad6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.329235] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1045.329235] env[62066]: value = "task-1156726" [ 1045.329235] env[62066]: _type = "Task" [ 1045.329235] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.337536] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.487672] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.487998] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82d9edf0-339d-4baf-ab1d-904300839b53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.498330] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1045.498330] env[62066]: value = "task-1156727" [ 1045.498330] env[62066]: _type = "Task" [ 1045.498330] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.507668] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156727, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.512450] env[62066]: INFO nova.compute.resource_tracker [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating resource usage from migration 61fabdd6-b55d-4ae7-930b-c6291b37608e [ 1045.548873] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156725, 'name': PowerOffVM_Task, 'duration_secs': 0.263136} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.549189] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.549365] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.549629] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fca584d-71a3-408c-a2da-ab9c4563c554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.665335] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc42428-a030-41ee-b54b-e24f6870777e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.676681] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff3e42b-2342-4fc1-8343-ee3cbfb0cc24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.712099] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97f4b3c-bfb0-4eac-87af-483f59f2e660 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.721654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35528bd6-80dc-4dda-8538-bf06e7a42b9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.738364] env[62066]: DEBUG nova.compute.provider_tree [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.840363] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156726, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.008435] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156727, 'name': PowerOffVM_Task, 'duration_secs': 0.47021} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.008713] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.008902] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1046.232194] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.232493] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.232735] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleting the datastore file [datastore2] 6e0065b0-da4f-4288-a858-8648572f8148 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.233074] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb02c270-b315-487c-9d65-3444d79b577b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.241096] env[62066]: DEBUG nova.scheduler.client.report [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.244601] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for the task: (returnval){ [ 1046.244601] env[62066]: value = "task-1156729" [ 1046.244601] env[62066]: _type = "Task" [ 1046.244601] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.253354] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.339900] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544239} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.340310] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.340506] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.340767] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc0d680d-872c-4366-a1ea-c5b74bb87bfa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.358921] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1046.358921] env[62066]: value = "task-1156730" [ 1046.358921] env[62066]: _type = "Task" [ 1046.358921] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.367737] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156730, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.514881] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.515326] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.515540] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.515751] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.515907] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.516179] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.516430] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.516600] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.516774] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.516946] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.517137] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.522527] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65c50771-63a3-49f0-92ac-4ffa7f7efba7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.539837] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1046.539837] env[62066]: value = "task-1156731" [ 1046.539837] env[62066]: _type = "Task" [ 1046.539837] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.550215] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156731, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.746256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.246s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.746680] env[62066]: INFO nova.compute.manager [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Migrating [ 1046.769934] env[62066]: DEBUG oslo_vmware.api [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Task: {'id': task-1156729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.770427] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.770693] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.770860] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.771052] env[62066]: INFO nova.compute.manager [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Took 1.75 seconds to destroy the instance on the hypervisor. [ 1046.771299] env[62066]: DEBUG oslo.service.loopingcall [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.771491] env[62066]: DEBUG nova.compute.manager [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.771584] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.869864] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156730, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188635} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.870181] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.870973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132f9647-c37d-4abd-85c4-ff581c12a1f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.893242] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.893557] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ecb7aee-6317-4cd1-a876-f820a85307ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.915375] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1046.915375] env[62066]: value = "task-1156732" [ 1046.915375] env[62066]: _type = "Task" [ 1046.915375] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.924089] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156732, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.050291] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156731, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.126182] env[62066]: DEBUG nova.compute.manager [req-8851dc84-7805-4c50-9351-6b15d50b9561 req-93ca9a62-888c-402c-9dad-9eaa321d5c2c service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Received event network-vif-deleted-00ffbc69-586f-4442-b0fc-616ec3c6c49b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.126182] env[62066]: INFO nova.compute.manager [req-8851dc84-7805-4c50-9351-6b15d50b9561 req-93ca9a62-888c-402c-9dad-9eaa321d5c2c service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Neutron deleted interface 00ffbc69-586f-4442-b0fc-616ec3c6c49b; detaching it from the instance and deleting it from the info cache [ 1047.126182] env[62066]: DEBUG nova.network.neutron [req-8851dc84-7805-4c50-9351-6b15d50b9561 req-93ca9a62-888c-402c-9dad-9eaa321d5c2c service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.265063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.265231] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.265396] env[62066]: DEBUG nova.network.neutron [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.427701] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156732, 'name': ReconfigVM_Task, 'duration_secs': 0.419183} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.428027] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.428657] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33cbe68f-cd7d-4a0d-946e-3db21172e4d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.436785] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1047.436785] env[62066]: value = "task-1156733" [ 1047.436785] env[62066]: _type = "Task" [ 1047.436785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.446666] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156733, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.549802] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156731, 'name': ReconfigVM_Task, 'duration_secs': 0.554613} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.550188] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1047.603991] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.628438] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-109b21a5-365c-45bc-93c1-72d0c5be87d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.637972] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe21fbe1-bd98-42f2-8bca-fe50e0ade289 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.667333] env[62066]: DEBUG nova.compute.manager [req-8851dc84-7805-4c50-9351-6b15d50b9561 req-93ca9a62-888c-402c-9dad-9eaa321d5c2c service nova] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Detach interface failed, port_id=00ffbc69-586f-4442-b0fc-616ec3c6c49b, reason: Instance 6e0065b0-da4f-4288-a858-8648572f8148 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1047.947980] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156733, 'name': Rename_Task, 'duration_secs': 0.141796} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.948362] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.948642] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c1f5931-c91a-47ab-99a7-8e5d54d1fe5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.955359] env[62066]: DEBUG nova.network.neutron [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [{"id": "65045fb9-30a1-493e-a710-cb21606de596", "address": "fa:16:3e:44:6a:ac", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65045fb9-30", "ovs_interfaceid": "65045fb9-30a1-493e-a710-cb21606de596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.957858] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1047.957858] env[62066]: value = "task-1156734" [ 1047.957858] env[62066]: _type = "Task" [ 1047.957858] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.965557] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.056970] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.057312] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.057483] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.057708] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.057891] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.058108] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.058368] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.058566] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.058772] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.058971] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.059215] env[62066]: DEBUG nova.virt.hardware [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.064713] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1048.065087] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b12b1a73-a435-4608-91aa-e8c8c391d6e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.084161] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1048.084161] env[62066]: value = "task-1156735" [ 1048.084161] env[62066]: _type = "Task" [ 1048.084161] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.092891] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.108151] env[62066]: INFO nova.compute.manager [-] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Took 1.34 seconds to deallocate network for instance. [ 1048.458866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.470391] env[62066]: DEBUG oslo_vmware.api [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156734, 'name': PowerOnVM_Task, 'duration_secs': 0.439853} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.470651] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.470853] env[62066]: INFO nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Took 7.85 seconds to spawn the instance on the hypervisor. [ 1048.471048] env[62066]: DEBUG nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.471853] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1692e9-c101-4380-aa2f-2141b27c1d1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.595176] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156735, 'name': ReconfigVM_Task, 'duration_secs': 0.167784} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.595480] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1048.596326] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7df36e8-c882-4b60-b283-24d0a97e3e4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.619259] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c/volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.620084] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.620327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.620537] env[62066]: DEBUG nova.objects.instance [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lazy-loading 'resources' on Instance uuid 6e0065b0-da4f-4288-a858-8648572f8148 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.621474] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c445b4a-8744-401e-9e33-663831e73bbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.643405] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1048.643405] env[62066]: value = "task-1156736" [ 1048.643405] env[62066]: _type = "Task" [ 1048.643405] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.652309] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.762212] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10baa904-2e74-4f98-b4a2-b703b054711e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.770491] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ca9903-9c38-4593-8b21-e7548b89ecc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.801763] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a6bb8a-3ed0-4f26-af8d-3a5730c347d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.810658] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed9ae72-b9db-4b36-9040-363760b5f555 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.825560] env[62066]: DEBUG nova.compute.provider_tree [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.992289] env[62066]: INFO nova.compute.manager [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Took 12.64 seconds to build instance. [ 1049.156170] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156736, 'name': ReconfigVM_Task, 'duration_secs': 0.280171} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.156580] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c/volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.156702] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.328074] env[62066]: DEBUG nova.scheduler.client.report [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.494339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5197b31c-e093-4986-891e-4e1988c95ce3 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.150s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.664242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880e4a47-af20-458c-957f-661d992a7de8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.685649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83faf18-4fe3-455d-96cb-9aab2ef34c3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.704607] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.773395] env[62066]: DEBUG nova.compute.manager [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.773730] env[62066]: DEBUG nova.compute.manager [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing instance network info cache due to event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1049.774110] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.774381] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.774648] env[62066]: DEBUG nova.network.neutron [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.833016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.212s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.852716] env[62066]: INFO nova.scheduler.client.report [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Deleted allocations for instance 6e0065b0-da4f-4288-a858-8648572f8148 [ 1049.977953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f003ac-2d85-46c5-9900-cd564b5cdf37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.002017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1050.361562] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba1d0e9d-d133-457a-b6cf-a634c431f047 tempest-AttachInterfacesTestJSON-25497475 tempest-AttachInterfacesTestJSON-25497475-project-member] Lock "6e0065b0-da4f-4288-a858-8648572f8148" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.348s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.506074] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.506408] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12ea8e12-5b52-4540-9345-4192751498c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.514659] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1050.514659] env[62066]: value = "task-1156737" [ 1050.514659] env[62066]: _type = "Task" [ 1050.514659] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.523959] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.542026] env[62066]: DEBUG nova.network.neutron [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updated VIF entry in instance network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.542413] env[62066]: DEBUG nova.network.neutron [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.026635] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156737, 'name': PowerOffVM_Task, 'duration_secs': 0.203218} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.027070] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.027346] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.044707] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ecf50ed-6267-442e-81af-ec74e17b3965 req-dfb393ca-0349-48f2-a1a1-81248918907f service nova] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.351098] env[62066]: DEBUG nova.network.neutron [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Port f557ea42-1811-49ae-90bf-31943c924e96 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1051.536024] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1051.536882] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1051.537207] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1051.537517] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1051.537808] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1051.538129] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.543391] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e556f01-6e47-48be-986e-3e3b5ab85346 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.562110] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1051.562110] env[62066]: value = "task-1156738" [ 1051.562110] env[62066]: _type = "Task" [ 1051.562110] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.572018] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156738, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.077269] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156738, 'name': ReconfigVM_Task, 'duration_secs': 0.249611} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.077269] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.374418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.374671] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.374918] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.583437] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.583791] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.583890] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.584094] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.584283] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.584458] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.584673] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.584839] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.585019] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.585384] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.585585] env[62066]: DEBUG nova.virt.hardware [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.591295] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1052.591566] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6efe7839-a707-49a4-aecd-42810bd37621 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.611902] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1052.611902] env[62066]: value = "task-1156739" [ 1052.611902] env[62066]: _type = "Task" [ 1052.611902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.621787] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156739, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.124145] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156739, 'name': ReconfigVM_Task, 'duration_secs': 0.287023} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.124461] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1053.125295] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186a789a-9215-4058-b8af-654d3558db78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.148504] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.149154] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02be2531-dbd0-4ecc-a075-2ea8689e8c2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.168331] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1053.168331] env[62066]: value = "task-1156740" [ 1053.168331] env[62066]: _type = "Task" [ 1053.168331] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.177565] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156740, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.411031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.411031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.411178] env[62066]: DEBUG nova.network.neutron [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.678654] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156740, 'name': ReconfigVM_Task, 'duration_secs': 0.279554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.679031] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6/968f6419-835e-49fe-b3a1-15cada4d22c6.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.679231] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1054.159369] env[62066]: DEBUG nova.network.neutron [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.186046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c334d1-65d3-41ee-af68-d616705c1322 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.211285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cff8d5-5143-4e4d-962c-ed48c117b7f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.227467] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1054.664023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.771009] env[62066]: DEBUG nova.network.neutron [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Port 65045fb9-30a1-493e-a710-cb21606de596 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1055.172611] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ba3b87-463b-450e-89ee-bec223d23b96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.183230] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6335c5f-7f67-4d19-8b18-43eb7bdd6e96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.716093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "f17d08b0-5ac0-442e-adba-fe7c91107206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.716537] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.798792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.799019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.799182] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.219214] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.324292] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a949ba2-8006-409a-ae70-34cc4c5cb571 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.350522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f08f929-42d1-4685-bdab-49f1974c14b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.363528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1056.740850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.741130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.742745] env[62066]: INFO nova.compute.claims [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.835105] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.835391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.835612] env[62066]: DEBUG nova.network.neutron [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1056.874230] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.874622] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2ecfd19-7672-46d0-bb66-75fce045162a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.882612] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1056.882612] env[62066]: value = "task-1156741" [ 1056.882612] env[62066]: _type = "Task" [ 1056.882612] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.890679] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.394548] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.541769] env[62066]: DEBUG nova.network.neutron [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [{"id": "65045fb9-30a1-493e-a710-cb21606de596", "address": "fa:16:3e:44:6a:ac", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65045fb9-30", "ovs_interfaceid": "65045fb9-30a1-493e-a710-cb21606de596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.889849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79610d6-3b30-4d5b-9591-41a5704a068f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.895536] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156741, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.900289] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dbc62a-0a74-446e-886c-e5d9fc952530 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.932088] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a8aa89-ef88-4833-b4eb-ecbb45a186c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.940160] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8462016f-703b-4715-8284-c626a06073cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.954687] env[62066]: DEBUG nova.compute.provider_tree [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.044227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.394791] env[62066]: DEBUG oslo_vmware.api [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156741, 'name': PowerOnVM_Task, 'duration_secs': 1.250742} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.395158] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1058.395432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeed9a-5e0f-4b9d-8298-72085f94c5b1 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.457558] env[62066]: DEBUG nova.scheduler.client.report [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.564488] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a006811-9c79-44c1-93d8-fe09ffe36411 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.583264] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e0aedd-77bb-4274-8b4c-219d4b4ca6a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.590178] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.962489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.963019] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.096992] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1059.097270] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28a0c062-398c-4248-8fe9-f7852ec6ea52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.106186] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1059.106186] env[62066]: value = "task-1156742" [ 1059.106186] env[62066]: _type = "Task" [ 1059.106186] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.113994] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.468035] env[62066]: DEBUG nova.compute.utils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1059.469461] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.469643] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.520990] env[62066]: DEBUG nova.policy [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a04ed71823074374b38358046b120dc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7794529140a64b9e910c6afa32dd7095', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.617037] env[62066]: DEBUG oslo_vmware.api [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156742, 'name': PowerOnVM_Task, 'duration_secs': 0.438701} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.617037] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.617037] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61b53b37-f16a-4b57-af6d-c50df85a85e9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance '968f6419-835e-49fe-b3a1-15cada4d22c6' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1059.800928] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Successfully created port: 6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.973589] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.657041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.657380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.657540] env[62066]: DEBUG nova.compute.manager [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Going to confirm migration 6 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1060.986876] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.037266] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.037266] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.039044] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.039044] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.039044] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.039044] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.039044] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.039483] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.039653] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.040043] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.040503] env[62066]: DEBUG nova.virt.hardware [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.042307] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6796e3-d37a-41e9-9597-a8e5cbd7ae96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.052756] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc254f29-b560-4347-ace8-250127eea650 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.219086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.219305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquired lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.219488] env[62066]: DEBUG nova.network.neutron [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1061.219673] env[62066]: DEBUG nova.objects.instance [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'info_cache' on Instance uuid f23700a0-c060-4f7e-8a7e-7d8acaeb7762 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.278058] env[62066]: DEBUG nova.compute.manager [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Received event network-vif-plugged-6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.278325] env[62066]: DEBUG oslo_concurrency.lockutils [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] Acquiring lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.278561] env[62066]: DEBUG oslo_concurrency.lockutils [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.278743] env[62066]: DEBUG oslo_concurrency.lockutils [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.278999] env[62066]: DEBUG nova.compute.manager [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] No waiting events found dispatching network-vif-plugged-6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.279317] env[62066]: WARNING nova.compute.manager [req-f96dd886-bf3a-4558-bd26-e81d1eb8ca77 req-08a16bae-9f96-4c64-bcc1-9cd2da45ee1e service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Received unexpected event network-vif-plugged-6c282951-0a14-441d-a734-a31ed6fffccd for instance with vm_state building and task_state spawning. [ 1061.410880] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Successfully updated port: 6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.672549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.672826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.673034] env[62066]: DEBUG nova.compute.manager [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Going to confirm migration 7 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1061.912887] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.913079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquired lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.913244] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.210051] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.210296] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquired lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.210442] env[62066]: DEBUG nova.network.neutron [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.210636] env[62066]: DEBUG nova.objects.instance [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'info_cache' on Instance uuid 968f6419-835e-49fe-b3a1-15cada4d22c6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.436084] env[62066]: DEBUG nova.network.neutron [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [{"id": "f557ea42-1811-49ae-90bf-31943c924e96", "address": "fa:16:3e:de:e9:ef", "network": {"id": "6afcc24a-d16e-434c-8141-4c0c6ab30fa6", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1130917606-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08737e043ede43b58298d1c1a834fa84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00b1e0dc-9aea-4ee2-a76b-1f0c3eaba916", "external-id": "nsx-vlan-transportzone-269", "segmentation_id": 269, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf557ea42-18", "ovs_interfaceid": "f557ea42-1811-49ae-90bf-31943c924e96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.450165] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1062.582761] env[62066]: DEBUG nova.network.neutron [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Updating instance_info_cache with network_info: [{"id": "6c282951-0a14-441d-a734-a31ed6fffccd", "address": "fa:16:3e:3c:6a:39", "network": {"id": "1601af4d-89cf-4eb7-b004-ac5f7bd488a1", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-691798628-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7794529140a64b9e910c6afa32dd7095", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c282951-0a", "ovs_interfaceid": "6c282951-0a14-441d-a734-a31ed6fffccd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.939338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Releasing lock "refresh_cache-f23700a0-c060-4f7e-8a7e-7d8acaeb7762" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.939625] env[62066]: DEBUG nova.objects.instance [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'migration_context' on Instance uuid f23700a0-c060-4f7e-8a7e-7d8acaeb7762 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.085927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Releasing lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.086203] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Instance network_info: |[{"id": "6c282951-0a14-441d-a734-a31ed6fffccd", "address": "fa:16:3e:3c:6a:39", "network": {"id": "1601af4d-89cf-4eb7-b004-ac5f7bd488a1", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-691798628-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7794529140a64b9e910c6afa32dd7095", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c282951-0a", "ovs_interfaceid": "6c282951-0a14-441d-a734-a31ed6fffccd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.086676] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:6a:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '01fe2e08-46f6-4cee-aefd-934461f8077d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c282951-0a14-441d-a734-a31ed6fffccd', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.094125] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Creating folder: Project (7794529140a64b9e910c6afa32dd7095). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1063.094393] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c81fddf-9dd2-4dcc-b97f-0f5a0f9d959e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.105477] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Created folder: Project (7794529140a64b9e910c6afa32dd7095) in parent group-v251573. [ 1063.105667] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Creating folder: Instances. Parent ref: group-v251780. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1063.105935] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-277a8c53-fc40-473a-bbcf-221cd77400c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.114802] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Created folder: Instances in parent group-v251780. [ 1063.115069] env[62066]: DEBUG oslo.service.loopingcall [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.115268] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.115495] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06655d32-342f-4c05-baab-45dbb1668792 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.132712] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.132712] env[62066]: value = "task-1156745" [ 1063.132712] env[62066]: _type = "Task" [ 1063.132712] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.143414] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156745, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.306637] env[62066]: DEBUG nova.compute.manager [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Received event network-changed-6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.306919] env[62066]: DEBUG nova.compute.manager [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Refreshing instance network info cache due to event network-changed-6c282951-0a14-441d-a734-a31ed6fffccd. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1063.307071] env[62066]: DEBUG oslo_concurrency.lockutils [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] Acquiring lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.307442] env[62066]: DEBUG oslo_concurrency.lockutils [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] Acquired lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.307442] env[62066]: DEBUG nova.network.neutron [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Refreshing network info cache for port 6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.419613] env[62066]: DEBUG nova.network.neutron [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [{"id": "65045fb9-30a1-493e-a710-cb21606de596", "address": "fa:16:3e:44:6a:ac", "network": {"id": "62948603-6fa1-4199-a9f9-572d8922ee25", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-133306603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c1a91ea6e0b4b2da6a16f327bc77a26", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65045fb9-30", "ovs_interfaceid": "65045fb9-30a1-493e-a710-cb21606de596", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.442384] env[62066]: DEBUG nova.objects.base [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1063.443641] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba13256c-259f-4c23-a1cd-5bc836867bcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.462920] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5ddd82-0fab-4cc0-af08-6cecfec4b56a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.469681] env[62066]: DEBUG oslo_vmware.api [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1063.469681] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5285b33a-05f9-1e71-c2ba-4898736a5761" [ 1063.469681] env[62066]: _type = "Task" [ 1063.469681] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.477640] env[62066]: DEBUG oslo_vmware.api [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5285b33a-05f9-1e71-c2ba-4898736a5761, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.643791] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156745, 'name': CreateVM_Task, 'duration_secs': 0.35192} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.643970] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.644670] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.644841] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.645183] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.645501] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c01ee8d6-edd1-4114-945a-68eebbbb01e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.649801] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1063.649801] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5274c3c3-433d-731f-29a0-940b73f267ef" [ 1063.649801] env[62066]: _type = "Task" [ 1063.649801] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.657162] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5274c3c3-433d-731f-29a0-940b73f267ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.922361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Releasing lock "refresh_cache-968f6419-835e-49fe-b3a1-15cada4d22c6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.922632] env[62066]: DEBUG nova.objects.instance [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'migration_context' on Instance uuid 968f6419-835e-49fe-b3a1-15cada4d22c6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.980447] env[62066]: DEBUG oslo_vmware.api [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5285b33a-05f9-1e71-c2ba-4898736a5761, 'name': SearchDatastore_Task, 'duration_secs': 0.007688} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.981588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.981588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.987572] env[62066]: DEBUG nova.network.neutron [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Updated VIF entry in instance network info cache for port 6c282951-0a14-441d-a734-a31ed6fffccd. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.987942] env[62066]: DEBUG nova.network.neutron [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Updating instance_info_cache with network_info: [{"id": "6c282951-0a14-441d-a734-a31ed6fffccd", "address": "fa:16:3e:3c:6a:39", "network": {"id": "1601af4d-89cf-4eb7-b004-ac5f7bd488a1", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-691798628-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7794529140a64b9e910c6afa32dd7095", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "01fe2e08-46f6-4cee-aefd-934461f8077d", "external-id": "nsx-vlan-transportzone-806", "segmentation_id": 806, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c282951-0a", "ovs_interfaceid": "6c282951-0a14-441d-a734-a31ed6fffccd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.160515] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5274c3c3-433d-731f-29a0-940b73f267ef, 'name': SearchDatastore_Task, 'duration_secs': 0.039542} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.160775] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.161025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.161273] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.161427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.161610] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.161866] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5f2cd16-1932-4863-87de-95813bfffb72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.170165] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.170346] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.171135] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-407d0955-9140-441d-97f5-c841509a2a5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.175984] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1064.175984] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52c43a84-934a-24dd-c63d-b909d068c141" [ 1064.175984] env[62066]: _type = "Task" [ 1064.175984] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.183133] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c43a84-934a-24dd-c63d-b909d068c141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.427592] env[62066]: DEBUG nova.objects.base [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Object Instance<968f6419-835e-49fe-b3a1-15cada4d22c6> lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1064.428662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd30ca7-b9df-468a-86fc-00f30f42c44d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.450639] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cdbde75-0cc5-4a18-acc1-232774c3039a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.456113] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1064.456113] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b010eb-2d68-cba9-2bc9-0b94b8aca404" [ 1064.456113] env[62066]: _type = "Task" [ 1064.456113] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.464081] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b010eb-2d68-cba9-2bc9-0b94b8aca404, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.490591] env[62066]: DEBUG oslo_concurrency.lockutils [req-894f033e-2775-40aa-9421-6a7ac436541a req-c4297bbb-1a25-4579-927f-200bb91940ca service nova] Releasing lock "refresh_cache-f17d08b0-5ac0-442e-adba-fe7c91107206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.597104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc6bb57-3c90-49a6-a15f-a74fd09174c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.604630] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9e06bb-cbdd-4f49-9b9e-65cfaa5a1fa5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.637377] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e61b183-4407-4c66-b5fd-c78bd052f3dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.641535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.641811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.646524] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d083d6-6503-4a27-80a0-f03d7f075711 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.661195] env[62066]: DEBUG nova.compute.provider_tree [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.685013] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52c43a84-934a-24dd-c63d-b909d068c141, 'name': SearchDatastore_Task, 'duration_secs': 0.015158} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.685731] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e5b0716-684d-4f12-8c91-00406644cf86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.690716] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1064.690716] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]525248ba-ff9e-d717-6803-8a6324e20ea4" [ 1064.690716] env[62066]: _type = "Task" [ 1064.690716] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.697778] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525248ba-ff9e-d717-6803-8a6324e20ea4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.967894] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b010eb-2d68-cba9-2bc9-0b94b8aca404, 'name': SearchDatastore_Task, 'duration_secs': 0.006541} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.968244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.147055] env[62066]: DEBUG nova.compute.utils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1065.163525] env[62066]: DEBUG nova.scheduler.client.report [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.201348] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]525248ba-ff9e-d717-6803-8a6324e20ea4, 'name': SearchDatastore_Task, 'duration_secs': 0.008599} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.202048] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.202048] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f17d08b0-5ac0-442e-adba-fe7c91107206/f17d08b0-5ac0-442e-adba-fe7c91107206.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.202217] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a50e0de5-a580-473e-b38b-b72007ca02cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.208926] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1065.208926] env[62066]: value = "task-1156746" [ 1065.208926] env[62066]: _type = "Task" [ 1065.208926] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.216483] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156746, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.650507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.720619] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156746, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.173351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.192s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.176517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.208s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.220222] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156746, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681383} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.220505] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] f17d08b0-5ac0-442e-adba-fe7c91107206/f17d08b0-5ac0-442e-adba-fe7c91107206.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.220737] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.220984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-296c7821-2f92-49cc-a7f4-643802e79f04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.227502] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1066.227502] env[62066]: value = "task-1156747" [ 1066.227502] env[62066]: _type = "Task" [ 1066.227502] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.235924] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.698537] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.698811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.699059] env[62066]: INFO nova.compute.manager [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Attaching volume c2a6c93c-d9cb-4b58-8ac9-c975d25fce52 to /dev/sdb [ 1066.700521] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.734095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37eebec-6bcf-4662-a7c4-715e4ef0af98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.740726] env[62066]: INFO nova.scheduler.client.report [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocation for migration 51f468c3-4a2a-4004-97ba-528e090fd6d3 [ 1066.744459] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.747472] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.748680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c85f1e7-1130-4030-bce8-dbdd6fa5f518 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.753544] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db533de6-43ac-489a-a029-67f5d33fbe25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.777304] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] f17d08b0-5ac0-442e-adba-fe7c91107206/f17d08b0-5ac0-442e-adba-fe7c91107206.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.778145] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed03843d-475c-46ff-a5c4-27e0d86d9bd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.799575] env[62066]: DEBUG nova.virt.block_device [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating existing volume attachment record: 31ec446d-e9ca-47f8-b202-f52defa10ef3 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1066.807701] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1066.807701] env[62066]: value = "task-1156748" [ 1066.807701] env[62066]: _type = "Task" [ 1066.807701] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.817961] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156748, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.863078] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb1cbdd-becb-4bc8-9bbf-945748003f7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.870038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb78d6f-5f59-497d-8467-4105fa5795e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.907242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b0845f-d69d-4848-85ce-6223886a6639 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.915328] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2a55ca-b164-4e10-bce4-f666849b8394 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.929928] env[62066]: DEBUG nova.compute.provider_tree [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.206200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.249551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dd490c1e-8de6-49ba-88b0-bdf195567b6a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.592s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.320782] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156748, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.433033] env[62066]: DEBUG nova.scheduler.client.report [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.683517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.683769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.819391] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156748, 'name': ReconfigVM_Task, 'duration_secs': 0.653988} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.819719] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Reconfigured VM instance instance-0000006e to attach disk [datastore2] f17d08b0-5ac0-442e-adba-fe7c91107206/f17d08b0-5ac0-442e-adba-fe7c91107206.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.820346] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0453196a-aeed-419f-8987-62d6c732ce62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.826333] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1067.826333] env[62066]: value = "task-1156752" [ 1067.826333] env[62066]: _type = "Task" [ 1067.826333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.833243] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156752, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.187027] env[62066]: DEBUG nova.compute.utils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1068.336522] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156752, 'name': Rename_Task, 'duration_secs': 0.12207} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.336817] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1068.337107] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cbff27d-7454-4657-843f-b53cc94d33a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.352837] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1068.352837] env[62066]: value = "task-1156753" [ 1068.352837] env[62066]: _type = "Task" [ 1068.352837] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.360217] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.444030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.267s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.447100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.241s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.447321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.447507] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1068.448636] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bf5dc5-45c8-42d6-bbfb-34c41e6ec367 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.457060] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073c8e44-0c02-40bf-b99c-7c4cd0d33563 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.471797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6827c1e3-80e6-4509-84ed-f0b7e75a5e01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.478098] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ceade5-fd6b-4cdc-9afe-0704a3b7e449 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.509901] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180319MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1068.510087] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.510310] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.689756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.862427] env[62066]: DEBUG oslo_vmware.api [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156753, 'name': PowerOnVM_Task, 'duration_secs': 0.452326} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.862692] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.862888] env[62066]: INFO nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Took 7.88 seconds to spawn the instance on the hypervisor. [ 1068.863087] env[62066]: DEBUG nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.863875] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc141a1-fb63-4337-82f4-625058bb2c59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.998877] env[62066]: INFO nova.scheduler.client.report [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocation for migration 61fabdd6-b55d-4ae7-930b-c6291b37608e [ 1069.272181] env[62066]: INFO nova.compute.manager [None req-d2b5455c-18df-438a-b229-ed63403c278a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Get console output [ 1069.272614] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-d2b5455c-18df-438a-b229-ed63403c278a tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] The console log is missing. Check your VSPC configuration [ 1069.378314] env[62066]: INFO nova.compute.manager [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Took 12.65 seconds to build instance. [ 1069.504915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.832s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.535876] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 5db50c22-048b-4cce-962a-3df1262f6e4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536036] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 6e76c799-9341-4315-8fa9-3c9acc1f8869 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536168] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 9c8bf6da-a21e-4566-9c15-4457b9a3de02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536287] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a451b1d1-73d3-41ad-b165-23983e48ace5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536405] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3852b155-6bdb-466d-af8f-0e3e665fe5c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536556] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f23700a0-c060-4f7e-8a7e-7d8acaeb7762 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536681] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 968f6419-835e-49fe-b3a1-15cada4d22c6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536795] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance f17d08b0-5ac0-442e-adba-fe7c91107206 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1069.536974] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1069.537122] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1069.627968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e84bb6-0dd7-41c5-a1f0-a62ca2f36c6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.635706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dba0a14-ab1b-41d3-a59a-ee9a2ab69eb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.666143] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996e1d0b-7400-430a-bfc8-40e21f1d8b83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.672649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b9c36b-7afe-4e69-95cb-e85a5bd3ce39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.685353] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.750456] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.750758] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.750937] env[62066]: INFO nova.compute.manager [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attaching volume d1efec0e-a3a1-445d-bd5c-1eccd180fe2a to /dev/sdb [ 1069.781028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55f2d99-303c-423b-a60f-72f5c2d5809e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.787486] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5373d68-9305-42cc-9d21-aaf217445e7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.800711] env[62066]: DEBUG nova.virt.block_device [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating existing volume attachment record: 443fb001-412a-4ce8-8160-bfef75a97c8a {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1069.880934] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbd4d122-ceca-489b-8b84-37ea8ba8cc80 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.164s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.188076] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.519061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "f17d08b0-5ac0-442e-adba-fe7c91107206" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.519061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.519463] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.519690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.519876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.522205] env[62066]: INFO nova.compute.manager [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Terminating instance [ 1070.524098] env[62066]: DEBUG nova.compute.manager [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.524309] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.525172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02113750-fdee-4bfe-8834-e2a002cf24df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.532762] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.532991] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c0931d4-1575-427d-9098-57ea19dcd002 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.539348] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1070.539348] env[62066]: value = "task-1156756" [ 1070.539348] env[62066]: _type = "Task" [ 1070.539348] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.548219] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.693331] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1070.693555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.183s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.693750] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.693888] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1070.700038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.700300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.700504] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.700710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.700883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.703185] env[62066]: INFO nova.compute.manager [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Terminating instance [ 1070.704864] env[62066]: DEBUG nova.compute.manager [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1070.705075] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.705902] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c5d849-e5c5-4034-9af5-b11ca260e69a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.713177] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.713409] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-786a327f-ef75-4729-bf0d-a3d3d2488dfa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.718935] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1070.718935] env[62066]: value = "task-1156757" [ 1070.718935] env[62066]: _type = "Task" [ 1070.718935] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.726626] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.048841] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156756, 'name': PowerOffVM_Task, 'duration_secs': 0.161928} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.049193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.049363] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.049550] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e0563b5-87dc-4a33-9a74-33c5d9a3aa72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.105467] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.105768] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.105962] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Deleting the datastore file [datastore2] f17d08b0-5ac0-442e-adba-fe7c91107206 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.106250] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eaf2991b-2c13-4cc5-b917-10ee1f26b1e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.112521] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for the task: (returnval){ [ 1071.112521] env[62066]: value = "task-1156759" [ 1071.112521] env[62066]: _type = "Task" [ 1071.112521] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.119834] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.211130] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] There are 52 instances to clean {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1071.211282] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: e273b275-2c03-4ac6-919e-46fddde91903] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.229094] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156757, 'name': PowerOffVM_Task, 'duration_secs': 0.221805} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.229360] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.229534] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.229792] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d78531dc-a0dd-436e-8717-e62b9e4ec097 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.289702] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.290152] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.290362] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleting the datastore file [datastore1] 968f6419-835e-49fe-b3a1-15cada4d22c6 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.290635] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cf1d524-e43e-401c-91a0-424b7ee1dc11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.297564] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for the task: (returnval){ [ 1071.297564] env[62066]: value = "task-1156761" [ 1071.297564] env[62066]: _type = "Task" [ 1071.297564] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.305261] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156761, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.347327] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1071.347639] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251784', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'name': 'volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a451b1d1-73d3-41ad-b165-23983e48ace5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'serial': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1071.349088] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4003b427-3ca4-4a28-9156-58f2452a2a37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.366721] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1498d867-330d-4e99-b149-44d4ec632a26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.391799] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52/volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.392155] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5b52caf-5dfd-4d21-96f9-e7162c777e2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.411097] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1071.411097] env[62066]: value = "task-1156762" [ 1071.411097] env[62066]: _type = "Task" [ 1071.411097] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.419718] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156762, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.625625] env[62066]: DEBUG oslo_vmware.api [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Task: {'id': task-1156759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183969} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.625848] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.626051] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.626233] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.626408] env[62066]: INFO nova.compute.manager [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1071.626686] env[62066]: DEBUG oslo.service.loopingcall [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.626875] env[62066]: DEBUG nova.compute.manager [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.626968] env[62066]: DEBUG nova.network.neutron [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.715023] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 08bc5e66-939c-436e-8b7b-dff7f8ff50f4] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.807900] env[62066]: DEBUG oslo_vmware.api [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Task: {'id': task-1156761, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.808215] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.808437] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.808643] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.808830] env[62066]: INFO nova.compute.manager [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1071.809116] env[62066]: DEBUG oslo.service.loopingcall [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.809330] env[62066]: DEBUG nova.compute.manager [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1071.809432] env[62066]: DEBUG nova.network.neutron [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1071.892546] env[62066]: DEBUG nova.compute.manager [req-cd50f4f6-6fbb-4346-b7f2-865629d601db req-74b770f8-4f3d-4b92-807d-b7bfcde63b66 service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Received event network-vif-deleted-6c282951-0a14-441d-a734-a31ed6fffccd {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.892851] env[62066]: INFO nova.compute.manager [req-cd50f4f6-6fbb-4346-b7f2-865629d601db req-74b770f8-4f3d-4b92-807d-b7bfcde63b66 service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Neutron deleted interface 6c282951-0a14-441d-a734-a31ed6fffccd; detaching it from the instance and deleting it from the info cache [ 1071.893127] env[62066]: DEBUG nova.network.neutron [req-cd50f4f6-6fbb-4346-b7f2-865629d601db req-74b770f8-4f3d-4b92-807d-b7bfcde63b66 service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.925059] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156762, 'name': ReconfigVM_Task, 'duration_secs': 0.336096} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.925059] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52/volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.934102] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-922a7074-ad13-48a4-8ebd-04796b6e71f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.950600] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1071.950600] env[62066]: value = "task-1156763" [ 1071.950600] env[62066]: _type = "Task" [ 1071.950600] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.958690] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156763, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.052098] env[62066]: DEBUG nova.compute.manager [req-7108c337-b203-46a3-a015-cb9053cb64cd req-d9e9f762-6bba-445a-9c9b-685830cba359 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Received event network-vif-deleted-65045fb9-30a1-493e-a710-cb21606de596 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.052344] env[62066]: INFO nova.compute.manager [req-7108c337-b203-46a3-a015-cb9053cb64cd req-d9e9f762-6bba-445a-9c9b-685830cba359 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Neutron deleted interface 65045fb9-30a1-493e-a710-cb21606de596; detaching it from the instance and deleting it from the info cache [ 1072.052597] env[62066]: DEBUG nova.network.neutron [req-7108c337-b203-46a3-a015-cb9053cb64cd req-d9e9f762-6bba-445a-9c9b-685830cba359 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.218470] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 7c8948df-1d0e-4334-8df2-b7f0bf5f5dbc] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.370340] env[62066]: DEBUG nova.network.neutron [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.396686] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6358354f-4159-4d5a-b1fe-af8640f49b32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.406218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e21231-3474-482d-b133-0327a3c8a250 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.434172] env[62066]: DEBUG nova.compute.manager [req-cd50f4f6-6fbb-4346-b7f2-865629d601db req-74b770f8-4f3d-4b92-807d-b7bfcde63b66 service nova] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Detach interface failed, port_id=6c282951-0a14-441d-a734-a31ed6fffccd, reason: Instance f17d08b0-5ac0-442e-adba-fe7c91107206 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1072.459694] env[62066]: DEBUG oslo_vmware.api [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156763, 'name': ReconfigVM_Task, 'duration_secs': 0.187939} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.460014] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251784', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'name': 'volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a451b1d1-73d3-41ad-b165-23983e48ace5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'serial': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1072.530704] env[62066]: DEBUG nova.network.neutron [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.555177] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43ebf0ad-b312-4a71-94dd-6973d8dfeb74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.564920] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898731d0-024e-40a2-9157-db753ec90a36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.592530] env[62066]: DEBUG nova.compute.manager [req-7108c337-b203-46a3-a015-cb9053cb64cd req-d9e9f762-6bba-445a-9c9b-685830cba359 service nova] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Detach interface failed, port_id=65045fb9-30a1-493e-a710-cb21606de596, reason: Instance 968f6419-835e-49fe-b3a1-15cada4d22c6 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1072.722176] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 7da117c1-0686-4b7e-9c33-5935a103dc3e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.873216] env[62066]: INFO nova.compute.manager [-] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Took 1.25 seconds to deallocate network for instance. [ 1073.033597] env[62066]: INFO nova.compute.manager [-] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Took 1.22 seconds to deallocate network for instance. [ 1073.225434] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 845748f7-d87b-4141-9932-235acd3d764c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.379564] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.379875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.380114] env[62066]: DEBUG nova.objects.instance [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lazy-loading 'resources' on Instance uuid f17d08b0-5ac0-442e-adba-fe7c91107206 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.494613] env[62066]: DEBUG nova.objects.instance [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.540483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.728791] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 045c2efa-e2fe-4d51-b267-e2b380fcf1d2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.983467] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab7274c-a764-470c-94e9-959dffd53c36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.992997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a52adc-f340-46ff-8e1e-acd866225657 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.025624] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfc71ee6-68bc-488b-905d-57ff63e44f6b tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.327s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.030026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d99a1b-c614-448a-a925-5e0b43a5ba64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.034579] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20d1fe2-70e7-42f5-836f-442647ede4fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.049173] env[62066]: DEBUG nova.compute.provider_tree [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.231826] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 10a2f245-93fd-4f6a-8c2f-897fd82269d3] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.344049] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1074.344316] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251785', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'name': 'volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'serial': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1074.345207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142f0f79-3458-46ab-adab-a623ee1ccf45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.362191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032d9450-238b-4ccd-8e22-ce37d61e96d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.385733] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a/volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.386419] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6474a386-77bf-4d8e-ac6e-5ded32029ac5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.403810] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1074.403810] env[62066]: value = "task-1156765" [ 1074.403810] env[62066]: _type = "Task" [ 1074.403810] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.411601] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.552163] env[62066]: DEBUG nova.scheduler.client.report [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.735888] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6e0065b0-da4f-4288-a858-8648572f8148] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.913124] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156765, 'name': ReconfigVM_Task, 'duration_secs': 0.367396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.913416] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfigured VM instance instance-00000069 to attach disk [datastore2] volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a/volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.918063] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-926836de-8682-4f26-ad69-996161fe84a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.931797] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1074.931797] env[62066]: value = "task-1156766" [ 1074.931797] env[62066]: _type = "Task" [ 1074.931797] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.938946] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.057889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.060291] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.520s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.060558] env[62066]: DEBUG nova.objects.instance [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lazy-loading 'resources' on Instance uuid 968f6419-835e-49fe-b3a1-15cada4d22c6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.076706] env[62066]: INFO nova.scheduler.client.report [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Deleted allocations for instance f17d08b0-5ac0-442e-adba-fe7c91107206 [ 1075.171029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.171189] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.239014] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 8bc2c334-ff46-44b2-b31a-d85273ec9a45] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.442394] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156766, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.583148] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b25ceb2-8e22-43aa-b1e6-dbe0812c3b77 tempest-ServerPasswordTestJSON-772526675 tempest-ServerPasswordTestJSON-772526675-project-member] Lock "f17d08b0-5ac0-442e-adba-fe7c91107206" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.064s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.673538] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1075.743017] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 16390d89-f2d1-4d6c-be62-b85e61906865] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.753771] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188d1f1c-00c7-437b-bfcb-fae72bb09906 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.763112] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dd98c8-959a-4358-bde1-8fb0bb38f299 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.795359] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab7f7f0-9029-4a83-ab8b-fba3d5c16160 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.802974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269135de-b4e9-4ff5-be8f-b78bcffc840e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.816660] env[62066]: DEBUG nova.compute.provider_tree [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.943738] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156766, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.192752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.246050] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ea7672f3-2ce2-4ef7-a202-8ce2c2ad8138] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.319571] env[62066]: DEBUG nova.scheduler.client.report [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.443365] env[62066]: DEBUG oslo_vmware.api [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156766, 'name': ReconfigVM_Task, 'duration_secs': 1.133018} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.443733] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251785', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'name': 'volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'serial': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1076.765287] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 2fc3eed6-7af2-4c34-a0d3-0498ae209f5f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.825254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.827744] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.635s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.829419] env[62066]: INFO nova.compute.claims [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.849561] env[62066]: INFO nova.scheduler.client.report [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Deleted allocations for instance 968f6419-835e-49fe-b3a1-15cada4d22c6 [ 1077.268101] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 4983ece5-bb4b-43cb-a928-324b2f31030e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.359457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-692c00b6-ad61-4a75-8b92-58072e834fd9 tempest-DeleteServersTestJSON-1162051174 tempest-DeleteServersTestJSON-1162051174-project-member] Lock "968f6419-835e-49fe-b3a1-15cada4d22c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.659s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.479839] env[62066]: DEBUG nova.objects.instance [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.771238] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f82c8bf2-24f3-4732-a56d-1bddaef14158] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.937018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4931ab77-0d75-46c1-8e64-e26681e9165d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.943442] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df58b016-b0a0-4580-a49f-ebb1c35b695c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.975223] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef84809-a4d5-4092-9d41-4fdaa41663c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.982187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee181c3-2bdf-423b-82aa-762e3b70cf9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.988293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8090f9d4-ff30-4797-aef2-cfecd105757f tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.238s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.996864] env[62066]: DEBUG nova.compute.provider_tree [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.275047] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d8aa9d38-0a2c-4036-9574-a85d2ed83fb2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.500500] env[62066]: DEBUG nova.scheduler.client.report [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.779036] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 00f44ecb-768a-4db3-b229-27bb6f27ede1] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.906666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.906915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.005594] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.006159] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.281861] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 7d7743d0-acf5-426e-b409-c00dca895048] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.410544] env[62066]: DEBUG nova.compute.utils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.511333] env[62066]: DEBUG nova.compute.utils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.512827] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.513085] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.558683] env[62066]: DEBUG nova.policy [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20873514f0fd4c4496d0cb5b2e500650', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a2bdc9041034d43b33453c202bd6cb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.785400] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: b2afea5a-2390-40b0-b860-c50f71ebac94] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.827981] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Successfully created port: 1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.913966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.016338] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.289023] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: cbf3c137-cc0f-42b7-96fb-2e1956e49b51] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.792031] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f578eda3-3d81-418b-bcb7-0a954835ed72] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.977071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.977246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.977382] env[62066]: INFO nova.compute.manager [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attaching volume c70602fe-7fea-4f49-8d75-9b092b5f1b0e to /dev/sdc [ 1081.007052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466c93b5-07e8-4fd0-b757-97a0ae34a62c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.013906] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f6fed1-445b-45c2-a625-1c56daaea519 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.028965] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.030736] env[62066]: DEBUG nova.virt.block_device [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating existing volume attachment record: 24c94ec8-0dff-417c-bbeb-4040a31cbd46 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1081.055576] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.055912] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.056103] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.056298] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.056553] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.056624] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.056829] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.057010] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.057401] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.057401] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.057529] env[62066]: DEBUG nova.virt.hardware [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.058327] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9543c43-41d1-44e4-bf2c-93c069d78333 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.065775] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96991260-0d95-4fab-9aee-5999cf4adc39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.295144] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 0fda5f59-55ac-4150-8402-00064d14c8ab] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.343235] env[62066]: DEBUG nova.compute.manager [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Received event network-vif-plugged-1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.343467] env[62066]: DEBUG oslo_concurrency.lockutils [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] Acquiring lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.343825] env[62066]: DEBUG oslo_concurrency.lockutils [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.343961] env[62066]: DEBUG oslo_concurrency.lockutils [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.344334] env[62066]: DEBUG nova.compute.manager [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] No waiting events found dispatching network-vif-plugged-1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.344431] env[62066]: WARNING nova.compute.manager [req-22498550-e3cf-4d2c-a1f1-30dceaaadcf0 req-c34b0fd2-e59e-49d6-9ca6-209dd37a139c service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Received unexpected event network-vif-plugged-1f4ea97b-6f76-4975-8105-ad7c465aa816 for instance with vm_state building and task_state spawning. [ 1081.432857] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Successfully updated port: 1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.799468] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 4914be5d-2cc8-48b7-96e5-9192e5c73fa3] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.935783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.935985] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.936138] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.302302] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 3b712d48-1365-4eed-8d6d-8ee6dbf51c2d] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.469066] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.605942] env[62066]: DEBUG nova.network.neutron [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updating instance_info_cache with network_info: [{"id": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "address": "fa:16:3e:2f:e3:85", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4ea97b-6f", "ovs_interfaceid": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.805021] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: da8b3bca-dbf5-4266-826d-ed93772fa04c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.109414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.109414] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Instance network_info: |[{"id": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "address": "fa:16:3e:2f:e3:85", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4ea97b-6f", "ovs_interfaceid": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.109644] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:e3:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bb10726-a946-47b9-b4b5-6916e3f14cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f4ea97b-6f76-4975-8105-ad7c465aa816', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.117382] env[62066]: DEBUG oslo.service.loopingcall [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.117523] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.118198] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66c3b477-ab73-40ed-9a91-c5966ff8e46b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.137792] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.137792] env[62066]: value = "task-1156769" [ 1083.137792] env[62066]: _type = "Task" [ 1083.137792] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.145211] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156769, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.308151] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 31e50f97-f873-44ad-9923-67923cdb8d3a] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.373588] env[62066]: DEBUG nova.compute.manager [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Received event network-changed-1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.373855] env[62066]: DEBUG nova.compute.manager [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Refreshing instance network info cache due to event network-changed-1f4ea97b-6f76-4975-8105-ad7c465aa816. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.374069] env[62066]: DEBUG oslo_concurrency.lockutils [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] Acquiring lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.374280] env[62066]: DEBUG oslo_concurrency.lockutils [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] Acquired lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.374391] env[62066]: DEBUG nova.network.neutron [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Refreshing network info cache for port 1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.648567] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156769, 'name': CreateVM_Task, 'duration_secs': 0.3237} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.648567] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.649200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.649363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.649695] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.649946] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-769213d3-c969-4441-8fdc-1ecd47f2ece9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.654298] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1083.654298] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52617550-4714-393b-b1e2-bd9cc13f524d" [ 1083.654298] env[62066]: _type = "Task" [ 1083.654298] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.662134] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52617550-4714-393b-b1e2-bd9cc13f524d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.812888] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f24f8067-07b2-4941-8464-c30b58983ec1] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.069948] env[62066]: DEBUG nova.network.neutron [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updated VIF entry in instance network info cache for port 1f4ea97b-6f76-4975-8105-ad7c465aa816. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1084.070344] env[62066]: DEBUG nova.network.neutron [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updating instance_info_cache with network_info: [{"id": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "address": "fa:16:3e:2f:e3:85", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4ea97b-6f", "ovs_interfaceid": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.165297] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52617550-4714-393b-b1e2-bd9cc13f524d, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.165600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.165891] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.166156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.166308] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.166498] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.166776] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac792f3a-c34c-468a-b610-94a14a27ce07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.174877] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.175061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.175766] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26e38c53-6cb1-449a-98a6-8f46efc4c06b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.180720] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1084.180720] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52029ae2-4575-c75a-e959-2197a5995dc9" [ 1084.180720] env[62066]: _type = "Task" [ 1084.180720] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.187738] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52029ae2-4575-c75a-e959-2197a5995dc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.314964] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 753f585a-22d7-4eeb-8580-4e3a68b5fd72] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.575937] env[62066]: DEBUG oslo_concurrency.lockutils [req-67a1db89-ed5c-476e-88e6-95856347b51f req-43c66f5b-9c41-4f8e-bfec-f81de83af1cd service nova] Releasing lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.691787] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52029ae2-4575-c75a-e959-2197a5995dc9, 'name': SearchDatastore_Task, 'duration_secs': 0.009003} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.692567] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aca5817e-196d-4e53-97f6-34aa04a9c02f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.700132] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1084.700132] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52abbb58-5234-4751-79d2-e6ad159fa108" [ 1084.700132] env[62066]: _type = "Task" [ 1084.700132] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.707545] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52abbb58-5234-4751-79d2-e6ad159fa108, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.818686] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6bd9da25-fcfc-41a1-b5d6-07d28ab37e15] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.210118] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52abbb58-5234-4751-79d2-e6ad159fa108, 'name': SearchDatastore_Task, 'duration_secs': 0.009914} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.210411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.210682] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4253e71f-4ca1-46ae-8c70-381dd9bc9383/4253e71f-4ca1-46ae-8c70-381dd9bc9383.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1085.210978] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01c52d00-6dba-496c-9edf-874a6c6a9fbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.217840] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1085.217840] env[62066]: value = "task-1156771" [ 1085.217840] env[62066]: _type = "Task" [ 1085.217840] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.226731] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.322631] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ecf9fd16-82c7-4bea-b6a9-7262e75effef] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.578379] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1085.578664] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251786', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'name': 'volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'serial': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1085.579659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb91f814-b6d9-4709-8d20-5dd3a9b66291 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.599270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b045a7f0-3dcf-4ef0-9e53-ae9378a5b8df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.628353] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e/volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.628543] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f0176d6-234e-4e2d-9fe4-165f26940a33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.650886] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1085.650886] env[62066]: value = "task-1156772" [ 1085.650886] env[62066]: _type = "Task" [ 1085.650886] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.659082] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.729887] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480097} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.730082] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 4253e71f-4ca1-46ae-8c70-381dd9bc9383/4253e71f-4ca1-46ae-8c70-381dd9bc9383.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.730315] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.730568] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8503b10d-9ed1-4274-bca2-1868431c55c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.736281] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1085.736281] env[62066]: value = "task-1156773" [ 1085.736281] env[62066]: _type = "Task" [ 1085.736281] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.745382] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156773, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.826207] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 3092f591-214f-40ba-a8e2-ccc6a0f007c9] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.163025] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156772, 'name': ReconfigVM_Task, 'duration_secs': 0.377725} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.163025] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfigured VM instance instance-00000069 to attach disk [datastore1] volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e/volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.167221] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec3ace5d-5536-49df-85b8-6b5a8e8f6ffb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.181708] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1086.181708] env[62066]: value = "task-1156774" [ 1086.181708] env[62066]: _type = "Task" [ 1086.181708] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.190724] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156774, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.246240] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156773, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.052708} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.246552] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.247435] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c73381-f31a-44b7-80ce-2f7050081565 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.273348] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 4253e71f-4ca1-46ae-8c70-381dd9bc9383/4253e71f-4ca1-46ae-8c70-381dd9bc9383.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.273877] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47e018de-7cce-4dde-bf3a-28dbc3d32c3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.298915] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1086.298915] env[62066]: value = "task-1156775" [ 1086.298915] env[62066]: _type = "Task" [ 1086.298915] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.309147] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.331886] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 68bf0e89-8236-44ef-b514-d25698eb23c3] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.691416] env[62066]: DEBUG oslo_vmware.api [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156774, 'name': ReconfigVM_Task, 'duration_secs': 0.164224} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.691744] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251786', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'name': 'volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'serial': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1086.732265] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.732594] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.812784] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156775, 'name': ReconfigVM_Task, 'duration_secs': 0.264694} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.812784] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 4253e71f-4ca1-46ae-8c70-381dd9bc9383/4253e71f-4ca1-46ae-8c70-381dd9bc9383.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.812969] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-016e8f8e-d09b-49a5-8dc3-125cf13d7fc3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.819195] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1086.819195] env[62066]: value = "task-1156776" [ 1086.819195] env[62066]: _type = "Task" [ 1086.819195] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.826902] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156776, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.835132] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5c1a4eb4-8f02-45d8-b09c-7b5cf25992d4] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.985566] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.985921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.986141] env[62066]: INFO nova.compute.manager [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Shelving [ 1087.235142] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1087.329792] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156776, 'name': Rename_Task, 'duration_secs': 0.141914} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.329792] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.329941] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13ee762c-b6af-4a00-91d5-b886f932570f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.336327] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1087.336327] env[62066]: value = "task-1156777" [ 1087.336327] env[62066]: _type = "Task" [ 1087.336327] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.340031] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 854a1ea4-515d-4a05-b179-22713b63f7c3] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.347149] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.494841] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.495145] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b81e5df-6c33-4c79-8fb9-5fb824a944e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.501954] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1087.501954] env[62066]: value = "task-1156778" [ 1087.501954] env[62066]: _type = "Task" [ 1087.501954] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.510270] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.731684] env[62066]: DEBUG nova.objects.instance [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.757522] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.757861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.759599] env[62066]: INFO nova.compute.claims [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.847051] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f3ef43e8-6092-44ac-8990-979810a9748f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.848039] env[62066]: DEBUG oslo_vmware.api [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156777, 'name': PowerOnVM_Task, 'duration_secs': 0.434307} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.848204] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.848411] env[62066]: INFO nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Took 6.82 seconds to spawn the instance on the hypervisor. [ 1087.848592] env[62066]: DEBUG nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.849564] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b6596a-49ef-4355-a3ec-99eaf1752eb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.011611] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156778, 'name': PowerOffVM_Task, 'duration_secs': 0.214857} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.011889] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1088.012690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc79316-227f-4ce3-af4d-875a913ef9e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.030842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72529d1e-45ff-4a66-9e4d-98d516906346 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.237221] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa801ee-5141-46cf-a03a-71f44709a1e1 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.260s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.349845] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f196adc6-86bb-4556-bf3d-c52b9108c000] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.365716] env[62066]: INFO nova.compute.manager [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Took 12.19 seconds to build instance. [ 1088.543350] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1088.543350] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-efe15628-e3ba-4f95-a83e-f025c27691aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.551978] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.552263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.553827] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1088.553827] env[62066]: value = "task-1156779" [ 1088.553827] env[62066]: _type = "Task" [ 1088.553827] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.561352] env[62066]: DEBUG nova.compute.manager [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Received event network-changed-1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.561576] env[62066]: DEBUG nova.compute.manager [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Refreshing instance network info cache due to event network-changed-1f4ea97b-6f76-4975-8105-ad7c465aa816. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.561864] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] Acquiring lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.561992] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] Acquired lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.562208] env[62066]: DEBUG nova.network.neutron [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Refreshing network info cache for port 1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.569572] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156779, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.853343] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 831c0bed-8a41-4672-aa08-d60b64f365b8] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.869555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d9d864-7c38-4601-aba0-29bc50a13421 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.698s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.872241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d237094-a08d-4ca4-99f0-7c1ee7a2af66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.880304] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4aa92e-cb92-49ea-b2ba-7bf73c478737 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.912987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c01b37e-1595-4a01-a5fa-126881119237 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.920324] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fc9259-64a8-4758-97e2-9cd84642f69d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.934328] env[62066]: DEBUG nova.compute.provider_tree [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.056811] env[62066]: INFO nova.compute.manager [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Detaching volume d1efec0e-a3a1-445d-bd5c-1eccd180fe2a [ 1089.072916] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156779, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.099841] env[62066]: INFO nova.virt.block_device [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attempting to driver detach volume d1efec0e-a3a1-445d-bd5c-1eccd180fe2a from mountpoint /dev/sdb [ 1089.100206] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.100417] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251785', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'name': 'volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'serial': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.105029] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46714946-2f18-4b6c-8874-b4be1848c958 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.131997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b373b4-2bae-42eb-aa20-11462697f240 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.143166] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b05afad-82c1-48a8-b7f3-5cb403ca0619 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.167332] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c7d40d-0721-47f4-a6dc-f04bf3a46f5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.185163] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] The volume has not been displaced from its original location: [datastore2] volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a/volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.190582] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.190877] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1018c60-fa59-4e92-a0cc-2ad36f7c7572 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.211579] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1089.211579] env[62066]: value = "task-1156780" [ 1089.211579] env[62066]: _type = "Task" [ 1089.211579] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.219363] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156780, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.296760] env[62066]: DEBUG nova.network.neutron [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updated VIF entry in instance network info cache for port 1f4ea97b-6f76-4975-8105-ad7c465aa816. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1089.297267] env[62066]: DEBUG nova.network.neutron [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updating instance_info_cache with network_info: [{"id": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "address": "fa:16:3e:2f:e3:85", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4ea97b-6f", "ovs_interfaceid": "1f4ea97b-6f76-4975-8105-ad7c465aa816", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.356555] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: b46b7795-b7f7-4c55-bc72-184bb7bb8842] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.436924] env[62066]: DEBUG nova.scheduler.client.report [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.569103] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156779, 'name': CreateSnapshot_Task, 'duration_secs': 0.634122} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.569335] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1089.570073] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8bc6a1-fa3d-4c43-8747-f723116f3a89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.721803] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156780, 'name': ReconfigVM_Task, 'duration_secs': 0.27269} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.721945] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1089.726505] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b1e6579-fb6e-4c2a-85f5-90a252d57a84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.740469] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1089.740469] env[62066]: value = "task-1156781" [ 1089.740469] env[62066]: _type = "Task" [ 1089.740469] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.747954] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.800926] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5c837a6-e372-4692-ad65-eee7834958dd req-5559799f-0ba1-4f48-8798-35bea75b1fb9 service nova] Releasing lock "refresh_cache-4253e71f-4ca1-46ae-8c70-381dd9bc9383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.860795] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 4723be94-3479-4e66-8088-914824c0e669] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.942428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.942958] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1090.088693] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1090.089030] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3f60cbe1-5539-40b9-b6a2-1373fa4452fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.097879] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1090.097879] env[62066]: value = "task-1156782" [ 1090.097879] env[62066]: _type = "Task" [ 1090.097879] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.106365] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156782, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.250023] env[62066]: DEBUG oslo_vmware.api [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156781, 'name': ReconfigVM_Task, 'duration_secs': 0.143131} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.250462] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251785', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'name': 'volume-d1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a', 'serial': 'd1efec0e-a3a1-445d-bd5c-1eccd180fe2a'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.363707] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 43d64f76-7f7a-4b95-b9df-c95218612998] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.448776] env[62066]: DEBUG nova.compute.utils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1090.450518] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1090.450751] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1090.499762] env[62066]: DEBUG nova.policy [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75a2f304461487e87c6d710eb7f830b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30c8aab967844a249c72080364b29278', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.609474] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156782, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.774469] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Successfully created port: a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1090.797082] env[62066]: DEBUG nova.objects.instance [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.867213] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: e95ee795-603e-4cbf-bcd6-1ba54b62a281] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.955150] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1091.108992] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156782, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.370935] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 0ba970e7-6b21-441a-81f7-2b4e7dfd4d76] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.609929] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156782, 'name': CloneVM_Task, 'duration_secs': 1.091077} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.610230] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Created linked-clone VM from snapshot [ 1091.610946] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5421ceb9-8ce5-4112-9bb5-431b5be45a8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.617863] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Uploading image 63bbc530-3455-441f-b33e-ea3b369e2743 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1091.642059] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1091.642059] env[62066]: value = "vm-251789" [ 1091.642059] env[62066]: _type = "VirtualMachine" [ 1091.642059] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1091.642325] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e772d341-f9c7-4749-923b-1566c4df37cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.650149] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease: (returnval){ [ 1091.650149] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aee034-3baa-ca47-3c27-1a06cb5deaf2" [ 1091.650149] env[62066]: _type = "HttpNfcLease" [ 1091.650149] env[62066]: } obtained for exporting VM: (result){ [ 1091.650149] env[62066]: value = "vm-251789" [ 1091.650149] env[62066]: _type = "VirtualMachine" [ 1091.650149] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1091.650375] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the lease: (returnval){ [ 1091.650375] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aee034-3baa-ca47-3c27-1a06cb5deaf2" [ 1091.650375] env[62066]: _type = "HttpNfcLease" [ 1091.650375] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1091.656390] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1091.656390] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aee034-3baa-ca47-3c27-1a06cb5deaf2" [ 1091.656390] env[62066]: _type = "HttpNfcLease" [ 1091.656390] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1091.798711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.805363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c3c7885c-c1ce-49ea-96d6-f997085fb5f6 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.253s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.806554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.008s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.874405] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d9163e64-309b-4381-8819-15757f83ac2e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.965961] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1091.988037] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.988360] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.988526] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.988718] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.988868] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.989033] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.989256] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.989420] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.989592] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.989760] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.989938] env[62066]: DEBUG nova.virt.hardware [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.990812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d93645-0f0f-4369-b8d0-f447e85e3db8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.999206] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32eb4aa4-3934-457f-9044-fe684e5405f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.153161] env[62066]: DEBUG nova.compute.manager [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Received event network-vif-plugged-a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.153413] env[62066]: DEBUG oslo_concurrency.lockutils [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.153635] env[62066]: DEBUG oslo_concurrency.lockutils [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.153856] env[62066]: DEBUG oslo_concurrency.lockutils [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.154079] env[62066]: DEBUG nova.compute.manager [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] No waiting events found dispatching network-vif-plugged-a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1092.154263] env[62066]: WARNING nova.compute.manager [req-d3c5fd3f-7fe2-4604-86f4-ff65b8ef9c81 req-6ce315c5-efeb-48f0-9bb6-f65081b6b7ea service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Received unexpected event network-vif-plugged-a30b3832-4a91-4ada-8984-d7cbc7d05969 for instance with vm_state building and task_state spawning. [ 1092.162701] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1092.162701] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aee034-3baa-ca47-3c27-1a06cb5deaf2" [ 1092.162701] env[62066]: _type = "HttpNfcLease" [ 1092.162701] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1092.162959] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1092.162959] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52aee034-3baa-ca47-3c27-1a06cb5deaf2" [ 1092.162959] env[62066]: _type = "HttpNfcLease" [ 1092.162959] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1092.163683] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2030b6d8-8d42-4ccc-b7f5-303775ddf74b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.170841] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1092.171032] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1092.234875] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Successfully updated port: a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.263058] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0c64ea99-1bed-4d97-a1ae-ccaa129263b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.309057] env[62066]: INFO nova.compute.manager [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Detaching volume c70602fe-7fea-4f49-8d75-9b092b5f1b0e [ 1092.346570] env[62066]: INFO nova.virt.block_device [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Attempting to driver detach volume c70602fe-7fea-4f49-8d75-9b092b5f1b0e from mountpoint /dev/sdc [ 1092.346945] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1092.347033] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251786', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'name': 'volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'serial': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1092.347952] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141fa2c8-ed81-4b2b-bd25-1b6d4b91dcca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.370897] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7428a1-3aa2-4fec-b1e5-bb71da6bf1f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.377632] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689247c1-14bc-4521-a43c-cc88ba1e133b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.380251] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a2251e56-2787-412d-89c9-eef111ee6d2b] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.409484] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ba261d-5eff-493b-932a-231ab2c4b960 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.424809] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] The volume has not been displaced from its original location: [datastore1] volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e/volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1092.430517] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfiguring VM instance instance-00000069 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.430930] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c028f25d-373d-435b-84ae-a81f07a4d491 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.452037] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1092.452037] env[62066]: value = "task-1156784" [ 1092.452037] env[62066]: _type = "Task" [ 1092.452037] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.459652] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.738042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.738303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.738560] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1092.883433] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: de9eb96d-e71b-4878-83b8-f75966dc3f48] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.964127] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156784, 'name': ReconfigVM_Task, 'duration_secs': 0.27713} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.964435] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Reconfigured VM instance instance-00000069 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.969309] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a686ec42-3be9-4d24-96c4-c019cc2d4019 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.984436] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1092.984436] env[62066]: value = "task-1156785" [ 1092.984436] env[62066]: _type = "Task" [ 1092.984436] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.992558] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.272882] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1093.386752] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 82e2a71a-d27f-4db9-8f84-16762d3d3bf6] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.422230] env[62066]: DEBUG nova.network.neutron [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.494620] env[62066]: DEBUG oslo_vmware.api [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156785, 'name': ReconfigVM_Task, 'duration_secs': 0.188414} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.495047] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251786', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'name': 'volume-c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c8bf6da-a21e-4566-9c15-4457b9a3de02', 'attached_at': '', 'detached_at': '', 'volume_id': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e', 'serial': 'c70602fe-7fea-4f49-8d75-9b092b5f1b0e'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1093.891140] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: ccb9f50f-dcc3-4d81-944e-d70803185ae1] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.924474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.924781] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Instance network_info: |[{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1093.925277] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:09:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a30b3832-4a91-4ada-8984-d7cbc7d05969', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.932975] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating folder: Project (30c8aab967844a249c72080364b29278). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1093.933561] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ba265a3-a1bc-4510-a72c-daddf3dba829 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.944787] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created folder: Project (30c8aab967844a249c72080364b29278) in parent group-v251573. [ 1093.945078] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating folder: Instances. Parent ref: group-v251790. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1093.945378] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-227f3442-0a1c-4063-99f8-7ad33bfeef80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.954134] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created folder: Instances in parent group-v251790. [ 1093.954373] env[62066]: DEBUG oslo.service.loopingcall [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.954563] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1093.954781] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-040dee09-d052-485f-909b-953c19a8c3c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.973757] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.973757] env[62066]: value = "task-1156788" [ 1093.973757] env[62066]: _type = "Task" [ 1093.973757] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.980969] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156788, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.039782] env[62066]: DEBUG nova.objects.instance [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'flavor' on Instance uuid 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.161094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.161465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.161660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.161955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.162266] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.164610] env[62066]: INFO nova.compute.manager [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Terminating instance [ 1094.166768] env[62066]: DEBUG nova.compute.manager [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1094.167030] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.167296] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-266b6358-d93d-4c17-b66e-a55fa39299a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.174718] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1094.174718] env[62066]: value = "task-1156789" [ 1094.174718] env[62066]: _type = "Task" [ 1094.174718] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.180477] env[62066]: DEBUG nova.compute.manager [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Received event network-changed-a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.180688] env[62066]: DEBUG nova.compute.manager [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Refreshing instance network info cache due to event network-changed-a30b3832-4a91-4ada-8984-d7cbc7d05969. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1094.180937] env[62066]: DEBUG oslo_concurrency.lockutils [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.181091] env[62066]: DEBUG oslo_concurrency.lockutils [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.181270] env[62066]: DEBUG nova.network.neutron [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Refreshing network info cache for port a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.188163] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.393863] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a8dd7483-0588-4f60-9504-20de799e69f1] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.484298] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156788, 'name': CreateVM_Task, 'duration_secs': 0.34897} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.484298] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1094.485072] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.486035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.486035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1094.486194] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e124675-e985-4ddc-8d34-c968032c1067 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.490667] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1094.490667] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e2fa1a-9f79-0e26-7dac-fe7430ccd98d" [ 1094.490667] env[62066]: _type = "Task" [ 1094.490667] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.498526] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e2fa1a-9f79-0e26-7dac-fe7430ccd98d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.688999] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156789, 'name': PowerOffVM_Task, 'duration_secs': 0.211521} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.689358] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.689564] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1094.689760] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251769', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'name': 'volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762', 'attached_at': '2024-10-07T01:01:33.000000', 'detached_at': '', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'serial': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1094.690854] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ef549d-c787-428a-8a99-950cba8522a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.709734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70df684a-5baa-4d8a-a416-d23579a3d0c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.716655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364014e5-22f5-42c8-9519-edfa97429ff6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.736188] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067129a0-2b28-4b55-a851-4fbbd502733e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.751273] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] The volume has not been displaced from its original location: [datastore2] volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c/volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1094.756560] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1094.759143] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33bd2891-1789-4ade-b51d-8856a0c50b4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.777261] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1094.777261] env[62066]: value = "task-1156790" [ 1094.777261] env[62066]: _type = "Task" [ 1094.777261] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.787019] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156790, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.899808] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d50f0505-66f4-412f-9744-25e0ea96277a] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.004020] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52e2fa1a-9f79-0e26-7dac-fe7430ccd98d, 'name': SearchDatastore_Task, 'duration_secs': 0.012714} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.004020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.004020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.004020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.004020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.004020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.004020] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15c0f883-b9aa-451d-8c26-0eae5f3ce865 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.015023] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.015023] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.015023] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30a2157e-548a-4586-8a11-5aaeac85cdea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.019621] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1095.019621] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52fdb0ef-045a-e183-dd96-8a1256d1419f" [ 1095.019621] env[62066]: _type = "Task" [ 1095.019621] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.023651] env[62066]: DEBUG nova.network.neutron [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updated VIF entry in instance network info cache for port a30b3832-4a91-4ada-8984-d7cbc7d05969. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.024164] env[62066]: DEBUG nova.network.neutron [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.030149] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fdb0ef-045a-e183-dd96-8a1256d1419f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.047222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7af4cb79-74d6-40d5-a66a-499c2e083011 tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.289587] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156790, 'name': ReconfigVM_Task, 'duration_secs': 0.204036} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.289940] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1095.294930] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1931be0-6851-4a7a-b488-666a64c2a342 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.310948] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1095.310948] env[62066]: value = "task-1156791" [ 1095.310948] env[62066]: _type = "Task" [ 1095.310948] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.322166] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156791, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.402252] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 26d87a85-0aa3-49b9-97ca-1b7fedbebb14] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.530505] env[62066]: DEBUG oslo_concurrency.lockutils [req-35e3bd28-de71-42b4-a98f-c42df15dc472 req-e58e9f84-72a7-4d9c-897b-8b1c30e34095 service nova] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.531062] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fdb0ef-045a-e183-dd96-8a1256d1419f, 'name': SearchDatastore_Task, 'duration_secs': 0.013963} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.531911] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54eec625-aba5-49ac-9b8d-a796e7758369 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.537362] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1095.537362] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52191e7b-fdd4-f02c-0862-0619eaf5996f" [ 1095.537362] env[62066]: _type = "Task" [ 1095.537362] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.545491] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52191e7b-fdd4-f02c-0862-0619eaf5996f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.821338] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156791, 'name': ReconfigVM_Task, 'duration_secs': 0.174148} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.821651] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251769', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'name': 'volume-6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'f23700a0-c060-4f7e-8a7e-7d8acaeb7762', 'attached_at': '2024-10-07T01:01:33.000000', 'detached_at': '', 'volume_id': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c', 'serial': '6e96e312-38b3-431f-aa8c-cfd4a71a9e9c'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1095.821927] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.822779] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d37f4f0-4165-4eb9-ad6d-03689db62909 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.830487] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.830736] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-003707c5-e3aa-4e69-b62f-11b0f441c4f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.897759] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.898013] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.898220] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore2] f23700a0-c060-4f7e-8a7e-7d8acaeb7762 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.898497] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b972b4bf-e8aa-418d-852a-9f08e495b480 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.905632] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1095.905632] env[62066]: value = "task-1156793" [ 1095.905632] env[62066]: _type = "Task" [ 1095.905632] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.906136] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 2f186d0a-91a0-4dc2-83bd-511099445af7] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.915372] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.048246] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52191e7b-fdd4-f02c-0862-0619eaf5996f, 'name': SearchDatastore_Task, 'duration_secs': 0.041501} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.048554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.048788] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1096.050043] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4476ccba-2503-4420-8d59-ab4505192c3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.055223] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1096.055223] env[62066]: value = "task-1156794" [ 1096.055223] env[62066]: _type = "Task" [ 1096.055223] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.063148] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.211083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.211264] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.211492] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.211712] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.211915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.214710] env[62066]: INFO nova.compute.manager [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Terminating instance [ 1096.216790] env[62066]: DEBUG nova.compute.manager [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.217012] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.217872] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bea5d6f-2763-49a6-a4e2-bc75cd00c821 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.226028] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.226281] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8556cbaa-df6c-4bdd-9fda-9e790be6d920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.232649] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1096.232649] env[62066]: value = "task-1156795" [ 1096.232649] env[62066]: _type = "Task" [ 1096.232649] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.240103] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.411610] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01e1df17-4b9d-4e12-bf6b-50b39c08bfbf] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.421259] env[62066]: DEBUG oslo_vmware.api [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140188} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.421633] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.421910] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.422203] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.422507] env[62066]: INFO nova.compute.manager [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1096.422836] env[62066]: DEBUG oslo.service.loopingcall [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.423126] env[62066]: DEBUG nova.compute.manager [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1096.423270] env[62066]: DEBUG nova.network.neutron [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1096.566996] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156794, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.744179] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156795, 'name': PowerOffVM_Task, 'duration_secs': 0.304084} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.744576] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1096.744700] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1096.744927] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01c4bd5a-f60a-44ea-8d0e-fa91ee5aae2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.810150] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.810405] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.810697] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleting the datastore file [datastore2] 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.810863] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd6c4d18-3e71-48fd-8078-33623d40db52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.818934] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for the task: (returnval){ [ 1096.818934] env[62066]: value = "task-1156797" [ 1096.818934] env[62066]: _type = "Task" [ 1096.818934] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.828217] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.915233] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 212dac6a-a291-4ca8-87fb-97ebcca7976c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.075937] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7443} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.076291] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1097.080018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1097.080018] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f5c16ad-2c1e-4461-8ad0-e2d6892bc9a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.083242] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1097.083242] env[62066]: value = "task-1156798" [ 1097.083242] env[62066]: _type = "Task" [ 1097.083242] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.091647] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.105411] env[62066]: DEBUG nova.compute.manager [req-fc29206d-16af-47f7-a785-b2117684ba21 req-e0e7958f-fc87-41e7-ba05-21e3d7821cde service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Received event network-vif-deleted-f557ea42-1811-49ae-90bf-31943c924e96 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.105630] env[62066]: INFO nova.compute.manager [req-fc29206d-16af-47f7-a785-b2117684ba21 req-e0e7958f-fc87-41e7-ba05-21e3d7821cde service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Neutron deleted interface f557ea42-1811-49ae-90bf-31943c924e96; detaching it from the instance and deleting it from the info cache [ 1097.105830] env[62066]: DEBUG nova.network.neutron [req-fc29206d-16af-47f7-a785-b2117684ba21 req-e0e7958f-fc87-41e7-ba05-21e3d7821cde service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.329188] env[62066]: DEBUG oslo_vmware.api [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Task: {'id': task-1156797, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.250944} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.329518] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.329774] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.330051] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.330266] env[62066]: INFO nova.compute.manager [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1097.330768] env[62066]: DEBUG oslo.service.loopingcall [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.331171] env[62066]: DEBUG nova.compute.manager [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.331282] env[62066]: DEBUG nova.network.neutron [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1097.419154] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.419538] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances with incomplete migration {{(pid=62066) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1097.582207] env[62066]: DEBUG nova.network.neutron [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.593273] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086326} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.593273] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.594048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a91e10d-49c6-4654-98a5-44433f527ac6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.619903] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.620278] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e42e916f-d6df-4ce3-8dfe-c02e6cf2bb58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.623538] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-679d4a8c-faf8-4f73-b08f-23a835d1307a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.652893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a093e1f5-46ef-4f99-a754-5c79e1fef501 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.662903] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1097.662903] env[62066]: value = "task-1156799" [ 1097.662903] env[62066]: _type = "Task" [ 1097.662903] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.672459] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156799, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.682650] env[62066]: DEBUG nova.compute.manager [req-fc29206d-16af-47f7-a785-b2117684ba21 req-e0e7958f-fc87-41e7-ba05-21e3d7821cde service nova] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Detach interface failed, port_id=f557ea42-1811-49ae-90bf-31943c924e96, reason: Instance f23700a0-c060-4f7e-8a7e-7d8acaeb7762 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1097.921930] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.086260] env[62066]: INFO nova.compute.manager [-] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Took 1.66 seconds to deallocate network for instance. [ 1098.174391] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156799, 'name': ReconfigVM_Task, 'duration_secs': 0.412537} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.174891] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.175823] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1be15f4e-8c20-4689-8666-50d2e7592ac3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.183726] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1098.183726] env[62066]: value = "task-1156800" [ 1098.183726] env[62066]: _type = "Task" [ 1098.183726] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.193372] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156800, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.278136] env[62066]: DEBUG nova.network.neutron [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.630787] env[62066]: INFO nova.compute.manager [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Took 0.54 seconds to detach 1 volumes for instance. [ 1098.633211] env[62066]: DEBUG nova.compute.manager [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Deleting volume: 6e96e312-38b3-431f-aa8c-cfd4a71a9e9c {{(pid=62066) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1098.694732] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156800, 'name': Rename_Task, 'duration_secs': 0.173141} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.694909] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.695213] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f15a532a-7bb0-482e-9883-53fe64e9d731 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.702564] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1098.702564] env[62066]: value = "task-1156802" [ 1098.702564] env[62066]: _type = "Task" [ 1098.702564] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.712085] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156802, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.780660] env[62066]: INFO nova.compute.manager [-] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Took 1.45 seconds to deallocate network for instance. [ 1099.131682] env[62066]: DEBUG nova.compute.manager [req-aa09ea3f-3e86-423c-b2ad-dd145488476d req-6b72c0df-84d8-4f26-b56f-adb334471482 service nova] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Received event network-vif-deleted-a8210a6e-c7ee-4503-b149-d547cf8b33b4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.173159] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.173548] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.173832] env[62066]: DEBUG nova.objects.instance [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'resources' on Instance uuid f23700a0-c060-4f7e-8a7e-7d8acaeb7762 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.216061] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156802, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.288094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.713775] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156802, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.798823] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0913fb8-1d80-428c-92a7-ebc1decce0f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.806833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214967ad-402a-4d51-9945-7255320ba182 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.838243] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6393e060-1a1c-44b4-b1bb-a14bf138d8cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.846024] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524c89b4-533a-401d-8c24-1868be20e994 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.859745] env[62066]: DEBUG nova.compute.provider_tree [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.213974] env[62066]: DEBUG oslo_vmware.api [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156802, 'name': PowerOnVM_Task, 'duration_secs': 1.276356} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.214361] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.214601] env[62066]: INFO nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Took 8.25 seconds to spawn the instance on the hypervisor. [ 1100.214789] env[62066]: DEBUG nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.215579] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b627fb-47ec-459b-be84-c4c5056f698c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.363125] env[62066]: DEBUG nova.scheduler.client.report [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.666544] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1100.667772] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b99ca7d-cac9-424b-bda6-d218a85b7fb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.673877] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1100.674077] env[62066]: ERROR oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk due to incomplete transfer. [ 1100.674321] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3d813f94-c325-4044-b1dc-b2d3bad657c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.680771] env[62066]: DEBUG oslo_vmware.rw_handles [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5242e1e5-0a41-f84e-eaeb-576e3ff66fd5/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1100.680968] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Uploaded image 63bbc530-3455-441f-b33e-ea3b369e2743 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1100.683336] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1100.683575] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7023f9a4-bdda-4c34-b29a-ae7e808575a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.688513] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1100.688513] env[62066]: value = "task-1156803" [ 1100.688513] env[62066]: _type = "Task" [ 1100.688513] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.695737] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156803, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.731944] env[62066]: INFO nova.compute.manager [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Took 12.99 seconds to build instance. [ 1100.868417] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.870999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.583s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.871333] env[62066]: DEBUG nova.objects.instance [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lazy-loading 'resources' on Instance uuid 9c8bf6da-a21e-4566-9c15-4457b9a3de02 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.889384] env[62066]: INFO nova.scheduler.client.report [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocations for instance f23700a0-c060-4f7e-8a7e-7d8acaeb7762 [ 1101.180079] env[62066]: DEBUG nova.compute.manager [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Received event network-changed-a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.180308] env[62066]: DEBUG nova.compute.manager [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Refreshing instance network info cache due to event network-changed-a30b3832-4a91-4ada-8984-d7cbc7d05969. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1101.180533] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.180680] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.180905] env[62066]: DEBUG nova.network.neutron [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Refreshing network info cache for port a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1101.199732] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156803, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.233404] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc68f657-2bce-46e3-8459-8665c4da7c2f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.501s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.397584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ae51afac-cd9e-4ac0-9a7b-2ca47f201a61 tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "f23700a0-c060-4f7e-8a7e-7d8acaeb7762" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.475649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f57edfd-b766-4ccc-ad9c-6c0f40965c5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.484195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c916fc4-6c80-4e2f-ad77-5ca9969bbdb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.516655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91c301c-31e2-49b9-90cb-7d94b2f2afc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.524913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebae95c-f583-401c-b2ab-ce3278c612bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.539297] env[62066]: DEBUG nova.compute.provider_tree [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.699289] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156803, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.893121] env[62066]: DEBUG nova.network.neutron [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updated VIF entry in instance network info cache for port a30b3832-4a91-4ada-8984-d7cbc7d05969. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1101.893510] env[62066]: DEBUG nova.network.neutron [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.042392] env[62066]: DEBUG nova.scheduler.client.report [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.066667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.066912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.067302] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.067518] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.067690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.069664] env[62066]: INFO nova.compute.manager [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Terminating instance [ 1102.071311] env[62066]: DEBUG nova.compute.manager [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.071510] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.072361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015c27cd-b54a-4aca-9741-4e13f0ff60a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.080291] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.080517] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edeaa4b0-6ad1-49b2-b5af-abe39e70b44a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.086697] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1102.086697] env[62066]: value = "task-1156804" [ 1102.086697] env[62066]: _type = "Task" [ 1102.086697] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.094275] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.203317] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156803, 'name': Destroy_Task, 'duration_secs': 1.30407} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.203695] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Destroyed the VM [ 1102.204054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1102.204440] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-97ce6369-8852-412d-a00d-097e0d099c1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.211366] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1102.211366] env[62066]: value = "task-1156805" [ 1102.211366] env[62066]: _type = "Task" [ 1102.211366] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.220267] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156805, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.396829] env[62066]: DEBUG oslo_concurrency.lockutils [req-8fe555fe-18f5-4671-b2bb-ccf1cd535795 req-40989f6f-ef30-41bd-b20c-3cdbd8e84e1a service nova] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.417733] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.417982] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.547822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.572135] env[62066]: INFO nova.scheduler.client.report [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Deleted allocations for instance 9c8bf6da-a21e-4566-9c15-4457b9a3de02 [ 1102.597046] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156804, 'name': PowerOffVM_Task, 'duration_secs': 0.213191} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.597402] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.597591] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.597869] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4fc45b5-6e0f-4266-986b-35d00777614c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.661908] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1102.662195] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1102.662409] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore2] 6e76c799-9341-4315-8fa9-3c9acc1f8869 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.662690] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45d4d4c6-f277-4a88-ae1c-420d9497e295 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.669356] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1102.669356] env[62066]: value = "task-1156807" [ 1102.669356] env[62066]: _type = "Task" [ 1102.669356] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.677816] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156807, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.720608] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156805, 'name': RemoveSnapshot_Task, 'duration_secs': 0.30631} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.720960] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1102.721320] env[62066]: DEBUG nova.compute.manager [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.722200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d22abf0-1709-4f52-bbdc-de1781f7f59e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.924023] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.924379] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1103.081066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e63fdb9-fbe4-41c3-bae0-a9a20196747a tempest-AttachVolumeTestJSON-564988674 tempest-AttachVolumeTestJSON-564988674-project-member] Lock "9c8bf6da-a21e-4566-9c15-4457b9a3de02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.869s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.179363] env[62066]: DEBUG oslo_vmware.api [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156807, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128839} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.179628] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.179847] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1103.180060] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1103.180251] env[62066]: INFO nova.compute.manager [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1103.180500] env[62066]: DEBUG oslo.service.loopingcall [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.180694] env[62066]: DEBUG nova.compute.manager [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.180789] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1103.234451] env[62066]: INFO nova.compute.manager [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Shelve offloading [ 1103.236178] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.236429] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a59bc74-aa81-447e-b64d-d8d86dfd288b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.242797] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1103.242797] env[62066]: value = "task-1156808" [ 1103.242797] env[62066]: _type = "Task" [ 1103.242797] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.250414] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.411296] env[62066]: DEBUG nova.compute.manager [req-e1f7bccc-24ea-4575-97ef-65c3280c8b88 req-09f40863-d5a3-4e64-b123-3d398a5f1169 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Received event network-vif-deleted-d4ccc7dd-b3cd-42b7-9363-9424832a0a2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.411499] env[62066]: INFO nova.compute.manager [req-e1f7bccc-24ea-4575-97ef-65c3280c8b88 req-09f40863-d5a3-4e64-b123-3d398a5f1169 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Neutron deleted interface d4ccc7dd-b3cd-42b7-9363-9424832a0a2d; detaching it from the instance and deleting it from the info cache [ 1103.411658] env[62066]: DEBUG nova.network.neutron [req-e1f7bccc-24ea-4575-97ef-65c3280c8b88 req-09f40863-d5a3-4e64-b123-3d398a5f1169 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.753568] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1103.753777] env[62066]: DEBUG nova.compute.manager [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.754604] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ba51ee-63a8-4fc0-b3a3-e5806ad90774 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.760528] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.760701] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.760878] env[62066]: DEBUG nova.network.neutron [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1103.884987] env[62066]: DEBUG nova.network.neutron [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.914586] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-243bb979-2cf6-4d0a-a192-86bc60da57ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.922938] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d553177e-629e-4767-8b5d-e6b473d53716 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.953651] env[62066]: DEBUG nova.compute.manager [req-e1f7bccc-24ea-4575-97ef-65c3280c8b88 req-09f40863-d5a3-4e64-b123-3d398a5f1169 service nova] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Detach interface failed, port_id=d4ccc7dd-b3cd-42b7-9363-9424832a0a2d, reason: Instance 6e76c799-9341-4315-8fa9-3c9acc1f8869 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1104.388396] env[62066]: INFO nova.compute.manager [-] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Took 1.21 seconds to deallocate network for instance. [ 1104.484091] env[62066]: DEBUG nova.network.neutron [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.894604] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.894888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.895135] env[62066]: DEBUG nova.objects.instance [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'resources' on Instance uuid 6e76c799-9341-4315-8fa9-3c9acc1f8869 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.986375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.345531] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.347112] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34331bb2-f476-4cfb-b98c-4b8dfdf95bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.354981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.355359] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-add9535f-c9f4-404f-bd9b-fa7cfc3dbdb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.420808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1105.421233] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1105.421525] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleting the datastore file [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.421826] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82451d85-6adf-4205-802d-c36ddf58bf02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.427919] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1105.427919] env[62066]: value = "task-1156811" [ 1105.427919] env[62066]: _type = "Task" [ 1105.427919] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.437939] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156811, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.440247] env[62066]: DEBUG nova.compute.manager [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-vif-unplugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.440486] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.440724] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.440916] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.441127] env[62066]: DEBUG nova.compute.manager [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] No waiting events found dispatching network-vif-unplugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.441330] env[62066]: WARNING nova.compute.manager [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received unexpected event network-vif-unplugged-50c2cd9e-175e-4926-86cb-f67b3875281f for instance with vm_state shelved and task_state shelving_offloading. [ 1105.441521] env[62066]: DEBUG nova.compute.manager [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.441709] env[62066]: DEBUG nova.compute.manager [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing instance network info cache due to event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1105.441931] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.442118] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.442312] env[62066]: DEBUG nova.network.neutron [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.493235] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f921d31-a5c6-44b0-ab2d-66637b786c0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.500617] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726cdb4f-650b-459d-95b8-4c515215286f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.533065] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9a0c90-05ca-4e10-bc82-6d68c706e47f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.540217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99e27f7-c46b-46f4-ae55-b89a6a28886c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.553249] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1105.938015] env[62066]: DEBUG oslo_vmware.api [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156811, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133036} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.938296] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.938509] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.938700] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.950323] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1105.950855] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951101] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951335] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951922] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951922] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951922] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1105.951922] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1105.957198] env[62066]: INFO nova.scheduler.client.report [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted allocations for instance 3852b155-6bdb-466d-af8f-0e3e665fe5c8 [ 1106.073664] env[62066]: ERROR nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [req-eac7d512-475b-420c-a9b5-47f9aad766ad] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-eac7d512-475b-420c-a9b5-47f9aad766ad"}]} [ 1106.090223] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1106.103057] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1106.103297] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1106.114526] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1106.132214] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1106.172631] env[62066]: DEBUG nova.network.neutron [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updated VIF entry in instance network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1106.173014] env[62066]: DEBUG nova.network.neutron [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": null, "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.198850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd640ccb-c425-416e-af4a-065548ee843c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.206568] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4623a30c-9115-4756-80ed-d9b9d566c674 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.238061] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cad9ac-9688-4836-ae4e-6e9524ed6f83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.245578] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60064777-1a98-48a3-ba81-8a3f80a6bfda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.259247] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1106.462343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.676350] env[62066]: DEBUG oslo_concurrency.lockutils [req-bc3c5d56-644d-43a7-866d-d1313d69309e req-ffc8308d-f2bd-4687-a382-bba6af91eea0 service nova] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.779763] env[62066]: ERROR nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [req-d03d7fc0-c8c9-4509-b81f-09dea67f7f22] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID cd4c0e36-9c88-4f73-a93c-1ff383ed97c4. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d03d7fc0-c8c9-4509-b81f-09dea67f7f22"}]} [ 1106.793463] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1106.805175] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1106.805436] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 163, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1106.815710] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1106.833389] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1106.914157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc18df0b-b5b6-4857-aff4-666a1c933ba1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.922352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f84200-b597-4bce-ac4d-99dac7fc9207 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.956052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09893247-79ed-498b-8de6-fd2b8598e9f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.963189] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f238807c-fca4-48bf-8f77-805cbe7b30a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.976173] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1107.505214] env[62066]: DEBUG nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updated inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with generation 152 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1107.505531] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 generation from 152 to 153 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1107.505674] env[62066]: DEBUG nova.compute.provider_tree [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1107.640467] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.010551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.115s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.013411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.551s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.013766] env[62066]: DEBUG nova.objects.instance [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'resources' on Instance uuid 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.027632] env[62066]: INFO nova.scheduler.client.report [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocations for instance 6e76c799-9341-4315-8fa9-3c9acc1f8869 [ 1108.516705] env[62066]: DEBUG nova.objects.instance [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'numa_topology' on Instance uuid 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.533953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-915f09d9-5871-4a90-aa26-0e0b3eacf9ea tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "6e76c799-9341-4315-8fa9-3c9acc1f8869" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.467s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.019508] env[62066]: DEBUG nova.objects.base [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Object Instance<3852b155-6bdb-466d-af8f-0e3e665fe5c8> lazy-loaded attributes: resources,numa_topology {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1109.082477] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9ac44f-6165-4045-bc23-e33507b82b84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.089733] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc764fb-c295-49f7-85be-5df466b622ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.118903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4135d1ca-2c9c-4eaa-8114-4fc34b34da7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.125745] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a96181d-d623-451c-a131-3932563380a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.139766] env[62066]: DEBUG nova.compute.provider_tree [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.642802] env[62066]: DEBUG nova.scheduler.client.report [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.148619] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.135s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.656727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-718187af-2b1b-4d57-b6e5-95f313f51264 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.670s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.657379] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.017s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.657487] env[62066]: INFO nova.compute.manager [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Unshelving [ 1110.852238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "5db50c22-048b-4cce-962a-3df1262f6e4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.852506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.852851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.852907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.853100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.855521] env[62066]: INFO nova.compute.manager [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Terminating instance [ 1110.857338] env[62066]: DEBUG nova.compute.manager [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1110.857535] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1110.858400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6fb5b1-9ecf-4b24-aa9c-de932f500529 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.866558] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.866782] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb0941d7-2fd7-45ed-9529-d5e968312050 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.872142] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1110.872142] env[62066]: value = "task-1156815" [ 1110.872142] env[62066]: _type = "Task" [ 1110.872142] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.881077] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.381766] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156815, 'name': PowerOffVM_Task, 'duration_secs': 0.181793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.381995] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.382191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.382435] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a7bcaef-8d38-4bdd-9c37-8557cf68db70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.440481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.440717] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.440930] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleting the datastore file [datastore1] 5db50c22-048b-4cce-962a-3df1262f6e4f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.441259] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe211ac3-d209-4476-b4ac-7f6fcb5f8206 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.447128] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for the task: (returnval){ [ 1111.447128] env[62066]: value = "task-1156817" [ 1111.447128] env[62066]: _type = "Task" [ 1111.447128] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.454552] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.681393] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.681722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.681939] env[62066]: DEBUG nova.objects.instance [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'pci_requests' on Instance uuid 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.957641] env[62066]: DEBUG oslo_vmware.api [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Task: {'id': task-1156817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13022} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.957923] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.958124] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1111.958358] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1111.958555] env[62066]: INFO nova.compute.manager [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1111.958798] env[62066]: DEBUG oslo.service.loopingcall [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.958990] env[62066]: DEBUG nova.compute.manager [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1111.959101] env[62066]: DEBUG nova.network.neutron [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1112.185782] env[62066]: DEBUG nova.objects.instance [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'numa_topology' on Instance uuid 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.187955] env[62066]: DEBUG nova.compute.manager [req-9a68b309-bd3e-4099-a311-9bd49944edcb req-18a0f3b4-e9c9-4e12-b10f-2a9242010d55 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Received event network-vif-deleted-05f9bc18-80b0-42b9-ba2b-8e552a0a77ab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.187955] env[62066]: INFO nova.compute.manager [req-9a68b309-bd3e-4099-a311-9bd49944edcb req-18a0f3b4-e9c9-4e12-b10f-2a9242010d55 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Neutron deleted interface 05f9bc18-80b0-42b9-ba2b-8e552a0a77ab; detaching it from the instance and deleting it from the info cache [ 1112.187955] env[62066]: DEBUG nova.network.neutron [req-9a68b309-bd3e-4099-a311-9bd49944edcb req-18a0f3b4-e9c9-4e12-b10f-2a9242010d55 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.664045] env[62066]: DEBUG nova.network.neutron [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.690979] env[62066]: INFO nova.compute.claims [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1112.693890] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1799c16f-d042-4bd0-8e72-47e078b94a4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.703170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fd612e-6a7a-4c4e-9495-d06f0ee5d480 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.727560] env[62066]: DEBUG nova.compute.manager [req-9a68b309-bd3e-4099-a311-9bd49944edcb req-18a0f3b4-e9c9-4e12-b10f-2a9242010d55 service nova] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Detach interface failed, port_id=05f9bc18-80b0-42b9-ba2b-8e552a0a77ab, reason: Instance 5db50c22-048b-4cce-962a-3df1262f6e4f could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1113.166073] env[62066]: INFO nova.compute.manager [-] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Took 1.21 seconds to deallocate network for instance. [ 1113.672961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.776609] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4afbb4-5e82-4eeb-ab5d-589ed98c4e32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.784088] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa27e5e6-7c35-4b04-85af-8bf9f504a1f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.814983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385051b6-7b2e-43b3-83a4-3660e8a07200 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.822079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d354b5-fc90-4cc1-9efa-09a2190803a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.835504] env[62066]: DEBUG nova.compute.provider_tree [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.342321] env[62066]: DEBUG nova.scheduler.client.report [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.848351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.166s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.850570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.178s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.850806] env[62066]: DEBUG nova.objects.instance [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lazy-loading 'resources' on Instance uuid 5db50c22-048b-4cce-962a-3df1262f6e4f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.878338] env[62066]: INFO nova.network.neutron [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating port 50c2cd9e-175e-4926-86cb-f67b3875281f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1115.437737] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff3836c-f310-41c1-9fcc-55f339dac1cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.448655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9b0ede-342d-468a-8aeb-166218e88b8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.496229] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ebbab7-0f0f-4dfe-b463-21152312fdf8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.504987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a29e204-748e-425f-b35e-185915a7473b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.519998] env[62066]: DEBUG nova.compute.provider_tree [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.022962] env[62066]: DEBUG nova.scheduler.client.report [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.280124] env[62066]: DEBUG nova.compute.manager [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.280376] env[62066]: DEBUG oslo_concurrency.lockutils [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.280590] env[62066]: DEBUG oslo_concurrency.lockutils [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.280769] env[62066]: DEBUG oslo_concurrency.lockutils [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.281171] env[62066]: DEBUG nova.compute.manager [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] No waiting events found dispatching network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1116.281410] env[62066]: WARNING nova.compute.manager [req-012a5f27-4610-44eb-82c6-9f2f7e53ca4a req-f3822c4d-767d-44ae-b8d9-dc609b71fb7b service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received unexpected event network-vif-plugged-50c2cd9e-175e-4926-86cb-f67b3875281f for instance with vm_state shelved_offloaded and task_state spawning. [ 1116.369321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.369515] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.369695] env[62066]: DEBUG nova.network.neutron [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.528191] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.551119] env[62066]: INFO nova.scheduler.client.report [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Deleted allocations for instance 5db50c22-048b-4cce-962a-3df1262f6e4f [ 1117.060036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5a9df02-faaf-498b-a854-cd7042eba1ec tempest-ServerActionsTestOtherA-1701747335 tempest-ServerActionsTestOtherA-1701747335-project-member] Lock "5db50c22-048b-4cce-962a-3df1262f6e4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.207s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.084991] env[62066]: DEBUG nova.network.neutron [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.587586] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.688751] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='166e79dc2d5437b19a28b33d1e945a9c',container_format='bare',created_at=2024-10-07T01:02:04Z,direct_url=,disk_format='vmdk',id=63bbc530-3455-441f-b33e-ea3b369e2743,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-508077974-shelved',owner='c468d5ba348d437f97a74e0da70bb42e',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-07T01:02:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.689033] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.689210] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.689401] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.689557] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.689714] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.690682] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.690682] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.690682] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.690682] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.690682] env[62066]: DEBUG nova.virt.hardware [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.691496] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240a0210-7129-470c-a8f5-78d9cc32b61e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.700063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6764ae1-400e-4815-8a1e-8ec64942a274 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.713788] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:b2:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '50c2cd9e-175e-4926-86cb-f67b3875281f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.721071] env[62066]: DEBUG oslo.service.loopingcall [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.721349] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1117.721563] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7984b04-1ee4-43f9-a958-e95192633f12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.740538] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.740538] env[62066]: value = "task-1156818" [ 1117.740538] env[62066]: _type = "Task" [ 1117.740538] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.748395] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156818, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.254104] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156818, 'name': CreateVM_Task, 'duration_secs': 0.286923} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.254104] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1118.254104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.254104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.254104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.254104] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1a9aeb3-d561-47e5-bc92-0a763d299f61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.258296] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1118.258296] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52856399-dbfe-aa34-a2d2-10552aeca3f9" [ 1118.258296] env[62066]: _type = "Task" [ 1118.258296] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.270749] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52856399-dbfe-aa34-a2d2-10552aeca3f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.317216] env[62066]: DEBUG nova.compute.manager [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.317822] env[62066]: DEBUG nova.compute.manager [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing instance network info cache due to event network-changed-50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1118.320043] env[62066]: DEBUG oslo_concurrency.lockutils [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] Acquiring lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.320043] env[62066]: DEBUG oslo_concurrency.lockutils [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] Acquired lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.320043] env[62066]: DEBUG nova.network.neutron [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Refreshing network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1118.769470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.769833] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Processing image 63bbc530-3455-441f-b33e-ea3b369e2743 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.769983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.770370] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquired lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.770370] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.770578] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de3d2472-f616-4ebd-909d-70e31653fe4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.791654] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.791868] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1118.792646] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ab945a-6051-44a2-ae11-64dedc091ae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.797953] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1118.797953] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]521267a5-15e4-84d9-5e60-257496477388" [ 1118.797953] env[62066]: _type = "Task" [ 1118.797953] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.806105] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]521267a5-15e4-84d9-5e60-257496477388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.164163] env[62066]: DEBUG nova.network.neutron [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updated VIF entry in instance network info cache for port 50c2cd9e-175e-4926-86cb-f67b3875281f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1119.164588] env[62066]: DEBUG nova.network.neutron [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [{"id": "50c2cd9e-175e-4926-86cb-f67b3875281f", "address": "fa:16:3e:7f:b2:20", "network": {"id": "0fb07cfd-07be-43ad-b9df-5194510fe462", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-206233062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c468d5ba348d437f97a74e0da70bb42e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50c2cd9e-17", "ovs_interfaceid": "50c2cd9e-175e-4926-86cb-f67b3875281f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.309760] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1119.310119] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Fetch image to [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a/OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1119.310489] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Downloading stream optimized image 63bbc530-3455-441f-b33e-ea3b369e2743 to [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a/OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a.vmdk on the data store datastore2 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1119.310489] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Downloading image file data 63bbc530-3455-441f-b33e-ea3b369e2743 to the ESX as VM named 'OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1119.384302] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1119.384302] env[62066]: value = "resgroup-9" [ 1119.384302] env[62066]: _type = "ResourcePool" [ 1119.384302] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1119.384653] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2f9916c3-8a09-432c-8132-f5043047cd63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.405197] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease: (returnval){ [ 1119.405197] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e61ee8-0f29-e1fe-a5ed-c0575d8050d8" [ 1119.405197] env[62066]: _type = "HttpNfcLease" [ 1119.405197] env[62066]: } obtained for vApp import into resource pool (val){ [ 1119.405197] env[62066]: value = "resgroup-9" [ 1119.405197] env[62066]: _type = "ResourcePool" [ 1119.405197] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1119.405472] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the lease: (returnval){ [ 1119.405472] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e61ee8-0f29-e1fe-a5ed-c0575d8050d8" [ 1119.405472] env[62066]: _type = "HttpNfcLease" [ 1119.405472] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1119.411821] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1119.411821] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e61ee8-0f29-e1fe-a5ed-c0575d8050d8" [ 1119.411821] env[62066]: _type = "HttpNfcLease" [ 1119.411821] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1119.667747] env[62066]: DEBUG oslo_concurrency.lockutils [req-df35898e-b5b4-4c4e-804d-ca9e7491d14d req-e9d77d6d-1c4f-4e16-ba89-76d84a6cde63 service nova] Releasing lock "refresh_cache-3852b155-6bdb-466d-af8f-0e3e665fe5c8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.914017] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1119.914017] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e61ee8-0f29-e1fe-a5ed-c0575d8050d8" [ 1119.914017] env[62066]: _type = "HttpNfcLease" [ 1119.914017] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1119.914341] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1119.914341] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52e61ee8-0f29-e1fe-a5ed-c0575d8050d8" [ 1119.914341] env[62066]: _type = "HttpNfcLease" [ 1119.914341] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1119.915115] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee5d3ce-ba25-4999-93b2-aa483adfd389 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.922356] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1119.922542] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1119.986366] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d9a4557c-d46a-4ef7-8809-6a9fb6ea6a07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.222221] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1121.222453] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1121.223843] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b48258-d256-4a56-8d4d-3b60b9ccbe4f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.232328] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1121.232328] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1121.232328] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-71649a66-f59b-4df1-803c-5879b79e4a37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.436372] env[62066]: DEBUG oslo_vmware.rw_handles [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52bdf482-75e3-97ab-8193-23c19da09996/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1121.436372] env[62066]: INFO nova.virt.vmwareapi.images [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Downloaded image file data 63bbc530-3455-441f-b33e-ea3b369e2743 [ 1121.437113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69efb138-ae02-4237-9ffb-fe9af44183b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.456644] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed06bdb4-5621-4849-a44a-6594ea6f3f5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.487893] env[62066]: INFO nova.virt.vmwareapi.images [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] The imported VM was unregistered [ 1121.489113] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1121.489507] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Creating directory with path [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.489904] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eeedd200-b4c0-4d59-adbd-0e0eb256016f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.501279] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Created directory with path [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.501760] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a/OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a.vmdk to [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1121.502190] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-48b639a7-09e1-4bed-bf75-10e9ba27428b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.509371] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1121.509371] env[62066]: value = "task-1156821" [ 1121.509371] env[62066]: _type = "Task" [ 1121.509371] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.517787] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.020813] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.189571] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "d139a247-e6bc-439e-b220-57cc5f135f04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.189833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.521014] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.692928] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1123.023046] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.215913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.216208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.217917] env[62066]: INFO nova.compute.claims [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1123.522495] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.021491] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156821, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.320773} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.021833] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a/OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a.vmdk to [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk. [ 1124.021970] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Cleaning up location [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1124.022166] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a0df333a-0a14-4038-bb81-ca859bc3a73a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.022431] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7e0d9e4-eed9-432c-bd37-db75d615bdb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.028927] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1124.028927] env[62066]: value = "task-1156822" [ 1124.028927] env[62066]: _type = "Task" [ 1124.028927] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.037154] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.290337] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce74ee6-d0c8-4a1d-bf42-fd7143f0adc3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.298984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8b8656-1ea3-430b-8f3c-91c97cbe2220 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.330767] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd080cde-3e2a-460b-a6e8-445d457bb3c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.337807] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d798a90f-92c2-4564-b961-51848d929d39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.350845] env[62066]: DEBUG nova.compute.provider_tree [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.538970] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033513} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.539332] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1124.539392] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Releasing lock "[datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.539605] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk to [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1124.539859] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72d95981-a799-4c8c-ae32-e1a82679919d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.545856] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1124.545856] env[62066]: value = "task-1156823" [ 1124.545856] env[62066]: _type = "Task" [ 1124.545856] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.553407] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.853829] env[62066]: DEBUG nova.scheduler.client.report [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.056104] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.359144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.143s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.359672] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1125.557153] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.865017] env[62066]: DEBUG nova.compute.utils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1125.866422] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1125.866607] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1125.908295] env[62066]: DEBUG nova.policy [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a295353365f04fc1b3ae93b1c8fde54b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f6c8792a8c9405a8c27cb170d23c251', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1126.057029] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.291245] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Successfully created port: 03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1126.369428] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1126.559277] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.581016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.581336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.581566] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.581777] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.582082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.584297] env[62066]: INFO nova.compute.manager [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Terminating instance [ 1126.586568] env[62066]: DEBUG nova.compute.manager [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1126.586841] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1126.587721] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfeffb81-646d-4eeb-9798-68a7002b3248 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.595273] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1126.595501] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86d7065c-aad8-42d8-aeea-6229701c9ec9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.602382] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1126.602382] env[62066]: value = "task-1156824" [ 1126.602382] env[62066]: _type = "Task" [ 1126.602382] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.610944] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.698624] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.058457] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156823, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.227592} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.058689] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/63bbc530-3455-441f-b33e-ea3b369e2743/63bbc530-3455-441f-b33e-ea3b369e2743.vmdk to [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1127.059481] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2248a25-064f-4b20-9c48-c7411751b52f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.081475] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.081777] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f9f65c0-aa9f-4755-9eda-adb3256ed082 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.100985] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1127.100985] env[62066]: value = "task-1156825" [ 1127.100985] env[62066]: _type = "Task" [ 1127.100985] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.111962] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.114914] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156824, 'name': PowerOffVM_Task, 'duration_secs': 0.187732} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.115187] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.115364] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1127.115601] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-522f4dde-de17-4666-aa5d-ea58808de84b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.173451] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1127.173791] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1127.174072] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleting the datastore file [datastore1] 4253e71f-4ca1-46ae-8c70-381dd9bc9383 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.174380] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-922b7ad6-ddf8-4110-b148-583700ef4ef1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.182101] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1127.182101] env[62066]: value = "task-1156827" [ 1127.182101] env[62066]: _type = "Task" [ 1127.182101] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.189680] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.201778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.202016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.202208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.202393] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1127.203204] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da98917-9fc2-4fa7-9bb0-2f98a3bf4fc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.211388] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58973d7a-6607-422d-8c5e-b80eed7555cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.225390] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7db5fda-0a8f-4bad-aca9-d60f1b4a0e8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.231608] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6966b258-900f-426d-a91a-6b83a01e6d9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.262620] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180883MB free_disk=163GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1127.262769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.262962] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.378713] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1127.404875] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.405185] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.405375] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.405581] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.405751] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.405914] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.406138] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.406317] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.406499] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.406719] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.406919] env[62066]: DEBUG nova.virt.hardware [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.407847] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9b9036-ea0a-4397-aec9-7db1a398b6f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.415493] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6148c8a1-b838-4ded-97fa-c1b88772dc6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.610788] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156825, 'name': ReconfigVM_Task, 'duration_secs': 0.319024} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.611090] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8/3852b155-6bdb-466d-af8f-0e3e665fe5c8.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.611721] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e39bb507-102d-4fdb-a584-9f2539f66b04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.617574] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1127.617574] env[62066]: value = "task-1156828" [ 1127.617574] env[62066]: _type = "Task" [ 1127.617574] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.625845] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156828, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.691922] env[62066]: DEBUG oslo_vmware.api [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274327} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.692303] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.692614] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1127.692788] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.693036] env[62066]: INFO nova.compute.manager [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1127.693349] env[62066]: DEBUG oslo.service.loopingcall [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.693596] env[62066]: DEBUG nova.compute.manager [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1127.693716] env[62066]: DEBUG nova.network.neutron [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1127.739564] env[62066]: DEBUG nova.compute.manager [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Received event network-vif-plugged-03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.739829] env[62066]: DEBUG oslo_concurrency.lockutils [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] Acquiring lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.740124] env[62066]: DEBUG oslo_concurrency.lockutils [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] Lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.740371] env[62066]: DEBUG oslo_concurrency.lockutils [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] Lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.740477] env[62066]: DEBUG nova.compute.manager [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] No waiting events found dispatching network-vif-plugged-03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1127.740668] env[62066]: WARNING nova.compute.manager [req-44cbee97-c863-41fa-a83c-0225f6add758 req-2f523dc3-d50b-403d-a7a1-a3bbb488b346 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Received unexpected event network-vif-plugged-03cdde9c-a1c4-478a-a539-89676b1cff6d for instance with vm_state building and task_state spawning. [ 1128.125919] env[62066]: DEBUG nova.compute.manager [req-53f335cf-b1f9-453e-a57e-6404cd220df6 req-08f3fada-c8f2-499a-8e92-7514e003a3c2 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Received event network-vif-deleted-1f4ea97b-6f76-4975-8105-ad7c465aa816 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.125919] env[62066]: INFO nova.compute.manager [req-53f335cf-b1f9-453e-a57e-6404cd220df6 req-08f3fada-c8f2-499a-8e92-7514e003a3c2 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Neutron deleted interface 1f4ea97b-6f76-4975-8105-ad7c465aa816; detaching it from the instance and deleting it from the info cache [ 1128.126420] env[62066]: DEBUG nova.network.neutron [req-53f335cf-b1f9-453e-a57e-6404cd220df6 req-08f3fada-c8f2-499a-8e92-7514e003a3c2 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.130393] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156828, 'name': Rename_Task, 'duration_secs': 0.14772} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.131012] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.131267] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a91dba0-7847-46e3-93a1-fd8d6503b8d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.140668] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1128.140668] env[62066]: value = "task-1156829" [ 1128.140668] env[62066]: _type = "Task" [ 1128.140668] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.154316] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.279535] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Successfully updated port: 03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1128.293290] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance a451b1d1-73d3-41ad-b165-23983e48ace5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.293473] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 4253e71f-4ca1-46ae-8c70-381dd9bc9383 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.293572] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01135457-b646-4c30-a730-1a18468ce58a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.293748] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3852b155-6bdb-466d-af8f-0e3e665fe5c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.293806] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance d139a247-e6bc-439e-b220-57cc5f135f04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1128.293993] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1128.294415] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1128.321182] env[62066]: DEBUG nova.compute.manager [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Received event network-changed-03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.321439] env[62066]: DEBUG nova.compute.manager [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Refreshing instance network info cache due to event network-changed-03cdde9c-a1c4-478a-a539-89676b1cff6d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1128.321703] env[62066]: DEBUG oslo_concurrency.lockutils [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] Acquiring lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.321830] env[62066]: DEBUG oslo_concurrency.lockutils [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] Acquired lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.322034] env[62066]: DEBUG nova.network.neutron [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Refreshing network info cache for port 03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1128.379646] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c974595-c51e-4ffa-9862-390614cde92b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.387323] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f967c4-2f17-48e6-9f01-a86a3f414951 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.421092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c2dd1b-6fe3-46ba-9cda-aec6ecd467fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.428550] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f76ef2-b43f-4770-90e3-5e3bb3207970 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.442122] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.604220] env[62066]: DEBUG nova.network.neutron [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.631958] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a640a988-0c9b-4307-802a-3b0e849116b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.641753] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632e8084-e495-4bf4-874b-9220612521f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.660693] env[62066]: DEBUG oslo_vmware.api [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156829, 'name': PowerOnVM_Task, 'duration_secs': 0.417331} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.661063] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.669803] env[62066]: DEBUG nova.compute.manager [req-53f335cf-b1f9-453e-a57e-6404cd220df6 req-08f3fada-c8f2-499a-8e92-7514e003a3c2 service nova] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Detach interface failed, port_id=1f4ea97b-6f76-4975-8105-ad7c465aa816, reason: Instance 4253e71f-4ca1-46ae-8c70-381dd9bc9383 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1128.760846] env[62066]: DEBUG nova.compute.manager [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.761166] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee64ac7-fca8-42ee-9ac8-a819f7e61a72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.782888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.851792] env[62066]: DEBUG nova.network.neutron [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1128.923046] env[62066]: DEBUG nova.network.neutron [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.945463] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.106498] env[62066]: INFO nova.compute.manager [-] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Took 1.41 seconds to deallocate network for instance. [ 1129.278833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-041f2bcb-a49c-4509-b954-568e60f416f2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.621s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.426140] env[62066]: DEBUG oslo_concurrency.lockutils [req-d5df9ddc-98e6-4a2d-a5a1-8be07be4f772 req-6ed7e1ba-7552-432f-839a-6660796ba609 service nova] Releasing lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.426912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquired lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.427094] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.449948] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1129.450173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.187s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.613295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.613595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.613816] env[62066]: DEBUG nova.objects.instance [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'resources' on Instance uuid 4253e71f-4ca1-46ae-8c70-381dd9bc9383 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.959334] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.088150] env[62066]: DEBUG nova.network.neutron [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Updating instance_info_cache with network_info: [{"id": "03cdde9c-a1c4-478a-a539-89676b1cff6d", "address": "fa:16:3e:78:10:e4", "network": {"id": "d8a35617-d70e-4eea-88ea-7edfff0ec99f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2047770689-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f6c8792a8c9405a8c27cb170d23c251", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ace50835-5731-4c77-b6c0-3076d7b4aa21", "external-id": "nsx-vlan-transportzone-270", "segmentation_id": 270, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03cdde9c-a1", "ovs_interfaceid": "03cdde9c-a1c4-478a-a539-89676b1cff6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.138041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.138041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.138366] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.138441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.138611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.140570] env[62066]: INFO nova.compute.manager [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Terminating instance [ 1130.142799] env[62066]: DEBUG nova.compute.manager [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1130.143010] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1130.143845] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e721168-0b6b-4408-aee3-97ee5736e8b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.155124] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1130.155375] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74e7d55b-1fb5-4bbd-92bf-75f9e41192f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.162353] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1130.162353] env[62066]: value = "task-1156830" [ 1130.162353] env[62066]: _type = "Task" [ 1130.162353] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.169867] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.191882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920689bd-3e2a-44a4-8681-720466709477 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.199152] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffc8236-b141-48f9-9e2c-7ec22aab8892 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.228651] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ac617a-0464-4d25-bd74-7617ee6ec71a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.236793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59029d24-2c4d-4256-9d14-07ab324a8255 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.250523] env[62066]: DEBUG nova.compute.provider_tree [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.590638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Releasing lock "refresh_cache-d139a247-e6bc-439e-b220-57cc5f135f04" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.590989] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance network_info: |[{"id": "03cdde9c-a1c4-478a-a539-89676b1cff6d", "address": "fa:16:3e:78:10:e4", "network": {"id": "d8a35617-d70e-4eea-88ea-7edfff0ec99f", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2047770689-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f6c8792a8c9405a8c27cb170d23c251", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ace50835-5731-4c77-b6c0-3076d7b4aa21", "external-id": "nsx-vlan-transportzone-270", "segmentation_id": 270, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03cdde9c-a1", "ovs_interfaceid": "03cdde9c-a1c4-478a-a539-89676b1cff6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1130.591470] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:10:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ace50835-5731-4c77-b6c0-3076d7b4aa21', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03cdde9c-a1c4-478a-a539-89676b1cff6d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.599711] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Creating folder: Project (3f6c8792a8c9405a8c27cb170d23c251). Parent ref: group-v251573. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1130.600017] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ed82584-adbc-4b12-a80c-3133f93258a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.612105] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Created folder: Project (3f6c8792a8c9405a8c27cb170d23c251) in parent group-v251573. [ 1130.612307] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Creating folder: Instances. Parent ref: group-v251795. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1130.612540] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c40e9b0-49b2-4f1a-943f-95565fa9a55c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.622774] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Created folder: Instances in parent group-v251795. [ 1130.622998] env[62066]: DEBUG oslo.service.loopingcall [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.623196] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1130.623388] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34742a50-d2a5-4c6d-bfcd-2aca7d9aff48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.640881] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.640881] env[62066]: value = "task-1156833" [ 1130.640881] env[62066]: _type = "Task" [ 1130.640881] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.647840] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156833, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.671951] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156830, 'name': PowerOffVM_Task, 'duration_secs': 0.199342} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.672350] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1130.672593] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1130.672887] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ccd4b17-a4b5-49a0-a6b6-f98acdf94597 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.733209] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1130.733475] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1130.733665] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleting the datastore file [datastore2] 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.733956] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f41ffb4-d341-4f82-8429-27146be59a90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.740735] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for the task: (returnval){ [ 1130.740735] env[62066]: value = "task-1156835" [ 1130.740735] env[62066]: _type = "Task" [ 1130.740735] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.750353] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156835, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.753491] env[62066]: DEBUG nova.scheduler.client.report [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.150427] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156833, 'name': CreateVM_Task, 'duration_secs': 0.307688} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.150674] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.151322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.151497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.151829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.152108] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abca33e9-b9bc-41a3-9287-7ca50abf130e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.156345] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1131.156345] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5246014c-a8ba-cdc7-6c57-4200c7c36ece" [ 1131.156345] env[62066]: _type = "Task" [ 1131.156345] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.163669] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5246014c-a8ba-cdc7-6c57-4200c7c36ece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.249659] env[62066]: DEBUG oslo_vmware.api [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Task: {'id': task-1156835, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149146} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.249933] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.250143] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1131.250328] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1131.250507] env[62066]: INFO nova.compute.manager [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1131.250744] env[62066]: DEBUG oslo.service.loopingcall [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.250942] env[62066]: DEBUG nova.compute.manager [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1131.251047] env[62066]: DEBUG nova.network.neutron [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1131.258440] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.276292] env[62066]: INFO nova.scheduler.client.report [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted allocations for instance 4253e71f-4ca1-46ae-8c70-381dd9bc9383 [ 1131.667370] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5246014c-a8ba-cdc7-6c57-4200c7c36ece, 'name': SearchDatastore_Task, 'duration_secs': 0.008977} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.667493] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.667722] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.668009] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.668180] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.668369] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.668634] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95e2abf4-a661-4736-9287-65c64c51637c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.677406] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.677598] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.678381] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6788a92-b690-466f-bc6a-da5a94dd041f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.683448] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1131.683448] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a51ae6-4338-2232-7d44-997b052bba53" [ 1131.683448] env[62066]: _type = "Task" [ 1131.683448] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.690949] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a51ae6-4338-2232-7d44-997b052bba53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.729956] env[62066]: DEBUG nova.compute.manager [req-b4299e88-b8b9-4702-8944-a3b9bd20258a req-76388fbe-6a87-4cab-8c8d-e3b30c728397 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Received event network-vif-deleted-50c2cd9e-175e-4926-86cb-f67b3875281f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.730190] env[62066]: INFO nova.compute.manager [req-b4299e88-b8b9-4702-8944-a3b9bd20258a req-76388fbe-6a87-4cab-8c8d-e3b30c728397 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Neutron deleted interface 50c2cd9e-175e-4926-86cb-f67b3875281f; detaching it from the instance and deleting it from the info cache [ 1131.730371] env[62066]: DEBUG nova.network.neutron [req-b4299e88-b8b9-4702-8944-a3b9bd20258a req-76388fbe-6a87-4cab-8c8d-e3b30c728397 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.785691] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f1e74839-e861-4c2f-93bb-b36e390dffe3 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "4253e71f-4ca1-46ae-8c70-381dd9bc9383" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.204s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.193959] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a51ae6-4338-2232-7d44-997b052bba53, 'name': SearchDatastore_Task, 'duration_secs': 0.009265} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.194853] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89186423-de32-48e5-8ce0-c9c9718b1642 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.200043] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1132.200043] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52b74a61-3eca-1515-7116-652658caed27" [ 1132.200043] env[62066]: _type = "Task" [ 1132.200043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.208987] env[62066]: DEBUG nova.network.neutron [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.210158] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b74a61-3eca-1515-7116-652658caed27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.232873] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35fe3e7e-7317-44e8-8682-6efe5c873e3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.242063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02e9d11-01fd-4f56-8a0f-0d3bd76d3297 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.267025] env[62066]: DEBUG nova.compute.manager [req-b4299e88-b8b9-4702-8944-a3b9bd20258a req-76388fbe-6a87-4cab-8c8d-e3b30c728397 service nova] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Detach interface failed, port_id=50c2cd9e-175e-4926-86cb-f67b3875281f, reason: Instance 3852b155-6bdb-466d-af8f-0e3e665fe5c8 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1132.525021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.525021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.710537] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52b74a61-3eca-1515-7116-652658caed27, 'name': SearchDatastore_Task, 'duration_secs': 0.01014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.710960] env[62066]: INFO nova.compute.manager [-] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Took 1.46 seconds to deallocate network for instance. [ 1132.711230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.711486] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d139a247-e6bc-439e-b220-57cc5f135f04/d139a247-e6bc-439e-b220-57cc5f135f04.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1132.713267] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-834a7539-7d0a-478e-9913-1213fdf73271 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.723189] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1132.723189] env[62066]: value = "task-1156836" [ 1132.723189] env[62066]: _type = "Task" [ 1132.723189] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.731423] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156836, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.028533] env[62066]: INFO nova.compute.manager [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Detaching volume c2a6c93c-d9cb-4b58-8ac9-c975d25fce52 [ 1133.067297] env[62066]: INFO nova.virt.block_device [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Attempting to driver detach volume c2a6c93c-d9cb-4b58-8ac9-c975d25fce52 from mountpoint /dev/sdb [ 1133.067637] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.068025] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251784', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'name': 'volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a451b1d1-73d3-41ad-b165-23983e48ace5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'serial': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.069114] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4f5cf7-b5d7-456a-9127-7483f1e9fc5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.092732] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbf1db3-9f59-45d4-932d-27ac820bcb50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.101982] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66384951-2348-4dab-ad66-282c55362f62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.123923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75b5f89-8d17-424d-a0b3-9629d4709666 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.141564] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] The volume has not been displaced from its original location: [datastore1] volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52/volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.146902] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.147679] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22d3edbe-65ed-40d1-82ae-31d82a1ccd07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.167176] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1133.167176] env[62066]: value = "task-1156837" [ 1133.167176] env[62066]: _type = "Task" [ 1133.167176] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.175552] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156837, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.220905] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.221293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.221503] env[62066]: DEBUG nova.objects.instance [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lazy-loading 'resources' on Instance uuid 3852b155-6bdb-466d-af8f-0e3e665fe5c8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.234093] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156836, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443996} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.234358] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] d139a247-e6bc-439e-b220-57cc5f135f04/d139a247-e6bc-439e-b220-57cc5f135f04.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.234574] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.234835] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0890672f-b6e5-4b40-a365-f75e162d657b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.241284] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1133.241284] env[62066]: value = "task-1156838" [ 1133.241284] env[62066]: _type = "Task" [ 1133.241284] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.249889] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156838, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.677234] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156837, 'name': ReconfigVM_Task, 'duration_secs': 0.244147} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.677452] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.681980] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ba2a4e9-b7ec-4f6f-a949-e4a39f0626ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.695733] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1133.695733] env[62066]: value = "task-1156839" [ 1133.695733] env[62066]: _type = "Task" [ 1133.695733] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.703057] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156839, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.752918] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156838, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078177} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.753282] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1133.754101] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed454c1-83f3-47f1-b817-90801d28d734 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.778272] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] d139a247-e6bc-439e-b220-57cc5f135f04/d139a247-e6bc-439e-b220-57cc5f135f04.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.780712] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30dcb77c-2502-42fb-9a25-da06caba7cff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.800459] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1133.800459] env[62066]: value = "task-1156840" [ 1133.800459] env[62066]: _type = "Task" [ 1133.800459] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.808376] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156840, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.821316] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a219e0-de4a-4cf9-bf92-647ab340f60b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.828303] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdd6131-2fa1-4267-b3b6-60c0b0f2ebd2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.859028] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18228358-d303-4bc5-bdf1-733963412eb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.865490] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e49574-6474-43f3-ac1c-7c0a4c5328be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.878555] env[62066]: DEBUG nova.compute.provider_tree [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.206090] env[62066]: DEBUG oslo_vmware.api [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156839, 'name': ReconfigVM_Task, 'duration_secs': 0.133092} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.206451] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251784', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'name': 'volume-c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a451b1d1-73d3-41ad-b165-23983e48ace5', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52', 'serial': 'c2a6c93c-d9cb-4b58-8ac9-c975d25fce52'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.310076] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156840, 'name': ReconfigVM_Task, 'duration_secs': 0.274945} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.310409] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Reconfigured VM instance instance-00000071 to attach disk [datastore2] d139a247-e6bc-439e-b220-57cc5f135f04/d139a247-e6bc-439e-b220-57cc5f135f04.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.310957] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d08098b-d84b-4341-bf08-4561546b4883 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.316817] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1134.316817] env[62066]: value = "task-1156841" [ 1134.316817] env[62066]: _type = "Task" [ 1134.316817] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.324237] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156841, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.381559] env[62066]: DEBUG nova.scheduler.client.report [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.446092] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.446309] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.446458] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1134.446576] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1134.745730] env[62066]: DEBUG nova.objects.instance [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.826618] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156841, 'name': Rename_Task, 'duration_secs': 0.129919} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.826934] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1134.827196] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0aded9e3-7efb-45f2-9d42-36400d6aacf3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.833447] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1134.833447] env[62066]: value = "task-1156842" [ 1134.833447] env[62066]: _type = "Task" [ 1134.833447] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.842850] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156842, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.887851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.909219] env[62066]: INFO nova.scheduler.client.report [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Deleted allocations for instance 3852b155-6bdb-466d-af8f-0e3e665fe5c8 [ 1134.950602] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1134.993561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.993750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.993873] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1134.994045] env[62066]: DEBUG nova.objects.instance [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lazy-loading 'info_cache' on Instance uuid a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.343232] env[62066]: DEBUG oslo_vmware.api [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156842, 'name': PowerOnVM_Task, 'duration_secs': 0.497474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.343627] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1135.343728] env[62066]: INFO nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Took 7.96 seconds to spawn the instance on the hypervisor. [ 1135.343919] env[62066]: DEBUG nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1135.344727] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c3cca2-d1d6-410f-8f06-7d4b5bc7cc63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.416709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebefacc5-c44b-4e17-8cce-c8859bcce4d2 tempest-ServerActionsTestOtherB-1860941823 tempest-ServerActionsTestOtherB-1860941823-project-member] Lock "3852b155-6bdb-466d-af8f-0e3e665fe5c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.279s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.754138] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e00ae389-bbb9-40eb-b0cb-e7f4296a38ec tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.229s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.861281] env[62066]: INFO nova.compute.manager [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Took 12.66 seconds to build instance. [ 1136.362318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d21f0a31-fe15-4270-a65e-c04bd5f73667 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.172s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.708814] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [{"id": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "address": "fa:16:3e:86:2c:ba", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcda4ad76-1c", "ovs_interfaceid": "cda4ad76-1cf1-4804-ac63-a7200bd27f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.768324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.768584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.768792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.768981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.769177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.771150] env[62066]: INFO nova.compute.manager [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Terminating instance [ 1136.772838] env[62066]: DEBUG nova.compute.manager [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1136.773048] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1136.773870] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c31c966-86cb-4096-b751-824051719fed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.781326] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1136.781543] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fc84aa9-0d96-4206-8853-be2f9fae786f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.788163] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1136.788163] env[62066]: value = "task-1156844" [ 1136.788163] env[62066]: _type = "Task" [ 1136.788163] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.795179] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156844, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.800624] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "d139a247-e6bc-439e-b220-57cc5f135f04" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.800845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.801051] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.801241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.801412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.803349] env[62066]: INFO nova.compute.manager [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Terminating instance [ 1136.805099] env[62066]: DEBUG nova.compute.manager [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1136.805304] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1136.806055] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305dec92-3c9c-45d5-ba8b-12fc5ff4da70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.812855] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1136.813088] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d23f532-8374-4228-8540-97bab0785355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.818500] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1136.818500] env[62066]: value = "task-1156845" [ 1136.818500] env[62066]: _type = "Task" [ 1136.818500] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.826613] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.211489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-a451b1d1-73d3-41ad-b165-23983e48ace5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.211713] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1137.211934] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.212110] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.212261] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.212407] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.212547] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.212685] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_power_states {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.299680] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156844, 'name': PowerOffVM_Task, 'duration_secs': 0.185806} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.300026] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.300251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.300532] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9075cedf-d0ea-43b9-aaa6-699b5571c8c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.327632] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156845, 'name': PowerOffVM_Task, 'duration_secs': 0.185021} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.327911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.328100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.328346] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-075ad903-dec4-4161-b8b4-c6a40d994c86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.357443] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.357707] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.357965] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleting the datastore file [datastore2] a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.358668] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d08636db-a4a1-4757-9501-da800cf0b6bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.364698] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1137.364698] env[62066]: value = "task-1156848" [ 1137.364698] env[62066]: _type = "Task" [ 1137.364698] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.372591] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.386787] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.387011] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.387226] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Deleting the datastore file [datastore2] d139a247-e6bc-439e-b220-57cc5f135f04 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.387503] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37978c4e-d1b6-4ccd-af17-46293e741734 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.393520] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for the task: (returnval){ [ 1137.393520] env[62066]: value = "task-1156849" [ 1137.393520] env[62066]: _type = "Task" [ 1137.393520] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.402088] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.716525] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Getting list of instances from cluster (obj){ [ 1137.716525] env[62066]: value = "domain-c8" [ 1137.716525] env[62066]: _type = "ClusterComputeResource" [ 1137.716525] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1137.717704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fec584-bff5-476e-b979-d0581a7b5230 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.727460] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Got total of 1 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1137.727609] env[62066]: WARNING nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] While synchronizing instance power states, found 3 instances in the database and 1 instances on the hypervisor. [ 1137.727812] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1137.727934] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1137.728104] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Triggering sync for uuid d139a247-e6bc-439e-b220-57cc5f135f04 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1137.728411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "a451b1d1-73d3-41ad-b165-23983e48ace5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.728649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.728845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.729133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "d139a247-e6bc-439e-b220-57cc5f135f04" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.729325] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.729458] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1137.730130] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3cd3ab-042c-4435-97b9-bb6d13cc0b58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.874246] env[62066]: DEBUG oslo_vmware.api [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13105} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.874526] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1137.874689] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1137.874893] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1137.875103] env[62066]: INFO nova.compute.manager [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1137.875348] env[62066]: DEBUG oslo.service.loopingcall [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1137.875541] env[62066]: DEBUG nova.compute.manager [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1137.875631] env[62066]: DEBUG nova.network.neutron [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1137.901722] env[62066]: DEBUG oslo_vmware.api [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Task: {'id': task-1156849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150959} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.901965] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1137.902164] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1137.902346] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1137.902522] env[62066]: INFO nova.compute.manager [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1137.902757] env[62066]: DEBUG oslo.service.loopingcall [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1137.902933] env[62066]: DEBUG nova.compute.manager [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1137.903038] env[62066]: DEBUG nova.network.neutron [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.206628] env[62066]: DEBUG nova.compute.manager [req-af5bda07-7f0c-4d81-9c79-11a1b2bf3ad1 req-cb50e655-433f-4c41-afc2-00e778a6e45f service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Received event network-vif-deleted-03cdde9c-a1c4-478a-a539-89676b1cff6d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.206863] env[62066]: INFO nova.compute.manager [req-af5bda07-7f0c-4d81-9c79-11a1b2bf3ad1 req-cb50e655-433f-4c41-afc2-00e778a6e45f service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Neutron deleted interface 03cdde9c-a1c4-478a-a539-89676b1cff6d; detaching it from the instance and deleting it from the info cache [ 1138.207061] env[62066]: DEBUG nova.network.neutron [req-af5bda07-7f0c-4d81-9c79-11a1b2bf3ad1 req-cb50e655-433f-4c41-afc2-00e778a6e45f service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.238082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.509s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.295506] env[62066]: DEBUG nova.compute.manager [req-c1411510-b3ed-4a17-96ba-eeaad3294d12 req-ed3da80e-10c0-4370-b4a4-00e94c6bc44a service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Received event network-vif-deleted-cda4ad76-1cf1-4804-ac63-a7200bd27f67 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.295717] env[62066]: INFO nova.compute.manager [req-c1411510-b3ed-4a17-96ba-eeaad3294d12 req-ed3da80e-10c0-4370-b4a4-00e94c6bc44a service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Neutron deleted interface cda4ad76-1cf1-4804-ac63-a7200bd27f67; detaching it from the instance and deleting it from the info cache [ 1138.295890] env[62066]: DEBUG nova.network.neutron [req-c1411510-b3ed-4a17-96ba-eeaad3294d12 req-ed3da80e-10c0-4370-b4a4-00e94c6bc44a service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.650322] env[62066]: DEBUG nova.network.neutron [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.713023] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e66b2ab-e5c5-48de-abfb-b462d5c6f566 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.720573] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a264c88-4c87-4176-a20b-d01fdab45748 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.745230] env[62066]: DEBUG nova.compute.manager [req-af5bda07-7f0c-4d81-9c79-11a1b2bf3ad1 req-cb50e655-433f-4c41-afc2-00e778a6e45f service nova] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Detach interface failed, port_id=03cdde9c-a1c4-478a-a539-89676b1cff6d, reason: Instance d139a247-e6bc-439e-b220-57cc5f135f04 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1138.775772] env[62066]: DEBUG nova.network.neutron [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.799737] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bca22055-b9b4-43ea-9fac-f000e819b31d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.809182] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522303c3-a1e8-405b-88d8-a04d8e1aad80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.834451] env[62066]: DEBUG nova.compute.manager [req-c1411510-b3ed-4a17-96ba-eeaad3294d12 req-ed3da80e-10c0-4370-b4a4-00e94c6bc44a service nova] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Detach interface failed, port_id=cda4ad76-1cf1-4804-ac63-a7200bd27f67, reason: Instance a451b1d1-73d3-41ad-b165-23983e48ace5 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1139.152924] env[62066]: INFO nova.compute.manager [-] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Took 1.25 seconds to deallocate network for instance. [ 1139.279574] env[62066]: INFO nova.compute.manager [-] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Took 1.40 seconds to deallocate network for instance. [ 1139.660475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.660475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.660475] env[62066]: DEBUG nova.objects.instance [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lazy-loading 'resources' on Instance uuid d139a247-e6bc-439e-b220-57cc5f135f04 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.785986] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.217619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fb6d28-f9c5-441c-aa8b-539da1f4172d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.225546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e379ce1-8b39-4f66-aed7-92e9fe85c384 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.256608] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833274a6-7653-4a88-9571-a5f42037f5cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.264595] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc0d659-e2a5-4871-86d0-1e565f0981cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.278091] env[62066]: DEBUG nova.compute.provider_tree [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.368025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.368025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.368025] env[62066]: DEBUG nova.compute.manager [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1140.368025] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7db71f0-64fe-40d0-8f2f-0c6a3fcbf505 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.374919] env[62066]: DEBUG nova.compute.manager [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1140.375507] env[62066]: DEBUG nova.objects.instance [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'flavor' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.781305] env[62066]: DEBUG nova.scheduler.client.report [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.882033] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1140.882280] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8773188-0107-4812-81f2-938bc0fa79c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.890158] env[62066]: DEBUG oslo_vmware.api [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1140.890158] env[62066]: value = "task-1156850" [ 1140.890158] env[62066]: _type = "Task" [ 1140.890158] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.898921] env[62066]: DEBUG oslo_vmware.api [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.287484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.290212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.504s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.290499] env[62066]: DEBUG nova.objects.instance [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'resources' on Instance uuid a451b1d1-73d3-41ad-b165-23983e48ace5 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.319054] env[62066]: INFO nova.scheduler.client.report [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Deleted allocations for instance d139a247-e6bc-439e-b220-57cc5f135f04 [ 1141.400036] env[62066]: DEBUG oslo_vmware.api [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156850, 'name': PowerOffVM_Task, 'duration_secs': 0.173441} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.400374] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1141.400562] env[62066]: DEBUG nova.compute.manager [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1141.401328] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87c77a9-9a1d-4ca7-a32a-36488280519d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.824631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d25918e7-9213-4061-8efd-00c45c341db5 tempest-ServerAddressesTestJSON-134512563 tempest-ServerAddressesTestJSON-134512563-project-member] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.024s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.826362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.096s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.826362] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] During sync_power_state the instance has a pending task (deleting). Skip. [ 1141.826362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "d139a247-e6bc-439e-b220-57cc5f135f04" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.831729] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32f7534-389c-4330-b1a6-b079ba813530 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.838952] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd87af6c-be9f-44b4-b28c-ffdcda844d48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.869815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d097c469-fc1f-40fb-8fbf-0671f047474d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.876883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67540395-2ead-468b-8414-ec298a62e1be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.890471] env[62066]: DEBUG nova.compute.provider_tree [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.913060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-45204e28-12d3-4a83-8b97-fc7eb0baf59b tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.393228] env[62066]: DEBUG nova.scheduler.client.report [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.734050] env[62066]: DEBUG nova.objects.instance [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'flavor' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.897966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.926178] env[62066]: INFO nova.scheduler.client.report [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted allocations for instance a451b1d1-73d3-41ad-b165-23983e48ace5 [ 1143.239473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.239658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.239843] env[62066]: DEBUG nova.network.neutron [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1143.240046] env[62066]: DEBUG nova.objects.instance [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'info_cache' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.435187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-06a7f287-947c-45bf-9ac1-72382f0a1e74 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.666s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.436428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.708s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.436729] env[62066]: INFO nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] During sync_power_state the instance has a pending task (deleting). Skip. [ 1143.437031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "a451b1d1-73d3-41ad-b165-23983e48ace5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.743390] env[62066]: DEBUG nova.objects.base [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Object Instance<01135457-b646-4c30-a730-1a18468ce58a> lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1144.498973] env[62066]: DEBUG nova.network.neutron [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.003731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.507392] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1145.507747] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be767eb0-1575-4250-be9a-8f93bd97992a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.515316] env[62066]: DEBUG oslo_vmware.api [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1145.515316] env[62066]: value = "task-1156851" [ 1145.515316] env[62066]: _type = "Task" [ 1145.515316] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.523175] env[62066]: DEBUG oslo_vmware.api [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.003658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.003945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.025081] env[62066]: DEBUG oslo_vmware.api [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156851, 'name': PowerOnVM_Task, 'duration_secs': 0.38219} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.025413] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1146.025648] env[62066]: DEBUG nova.compute.manager [None req-08111f37-a126-42e8-8457-f7ced2ab1a39 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1146.026418] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a4ee87-2ed4-4038-a506-7b1b4a22dfe9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.506450] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1147.027745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.028015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.029521] env[62066]: INFO nova.compute.claims [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1148.099811] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bbc9a6-8b79-48c4-9b52-c6bd3267ce0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.107578] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d821d1a2-cdde-4676-88bc-c3745dd2fb0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.137892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197ae51c-852b-4642-962b-f4c711b17bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.144983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662729e6-f74d-4202-a898-a3ccdb6804ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.157767] env[62066]: DEBUG nova.compute.provider_tree [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.661142] env[62066]: DEBUG nova.scheduler.client.report [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.166377] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.138s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.166872] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1149.672012] env[62066]: DEBUG nova.compute.utils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1149.673449] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1149.673620] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1149.720165] env[62066]: DEBUG nova.policy [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20873514f0fd4c4496d0cb5b2e500650', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a2bdc9041034d43b33453c202bd6cb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1149.982354] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Successfully created port: ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1150.176946] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1151.187299] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1151.215405] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1151.215816] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1151.216108] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1151.216435] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1151.216692] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1151.216955] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1151.217230] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1151.217409] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1151.217586] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1151.217753] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1151.217928] env[62066]: DEBUG nova.virt.hardware [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.218812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26437a4-68ed-412a-9dea-6224f5db3eed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.226586] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81aa5da-a56c-4795-bddf-a415e456fe1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.403267] env[62066]: DEBUG nova.compute.manager [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Received event network-vif-plugged-ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1151.403574] env[62066]: DEBUG oslo_concurrency.lockutils [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.403824] env[62066]: DEBUG oslo_concurrency.lockutils [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] Lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.404008] env[62066]: DEBUG oslo_concurrency.lockutils [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] Lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.404242] env[62066]: DEBUG nova.compute.manager [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] No waiting events found dispatching network-vif-plugged-ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1151.404417] env[62066]: WARNING nova.compute.manager [req-68fd9685-a95e-4957-9f03-9feedd0c6e01 req-a135d838-509b-4dd9-890d-8ebb2213f8b0 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Received unexpected event network-vif-plugged-ca7a22d5-2b52-42fe-a439-1ae3808aa672 for instance with vm_state building and task_state spawning. [ 1151.433889] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Successfully updated port: ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1151.937092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.937092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.937274] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1152.467684] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1152.588599] env[62066]: DEBUG nova.network.neutron [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating instance_info_cache with network_info: [{"id": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "address": "fa:16:3e:ce:29:d2", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca7a22d5-2b", "ovs_interfaceid": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.091281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.091721] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Instance network_info: |[{"id": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "address": "fa:16:3e:ce:29:d2", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca7a22d5-2b", "ovs_interfaceid": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1153.092211] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:29:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bb10726-a946-47b9-b4b5-6916e3f14cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca7a22d5-2b52-42fe-a439-1ae3808aa672', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1153.104095] env[62066]: DEBUG oslo.service.loopingcall [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.104430] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1153.104790] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80af2b99-a083-4093-8216-e15baecee737 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.137827] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1153.137827] env[62066]: value = "task-1156852" [ 1153.137827] env[62066]: _type = "Task" [ 1153.137827] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.146635] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156852, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.435660] env[62066]: DEBUG nova.compute.manager [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Received event network-changed-ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1153.435874] env[62066]: DEBUG nova.compute.manager [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Refreshing instance network info cache due to event network-changed-ca7a22d5-2b52-42fe-a439-1ae3808aa672. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1153.436112] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] Acquiring lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.436262] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] Acquired lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.436429] env[62066]: DEBUG nova.network.neutron [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Refreshing network info cache for port ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1153.649223] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156852, 'name': CreateVM_Task, 'duration_secs': 0.299619} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.649584] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1153.650015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.650193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.650510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1153.650750] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed412d11-a69a-458b-8695-87220ac3ba96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.654897] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1153.654897] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5234377f-30cd-32fa-5d53-63be7461af0b" [ 1153.654897] env[62066]: _type = "Task" [ 1153.654897] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.662168] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5234377f-30cd-32fa-5d53-63be7461af0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.154486] env[62066]: DEBUG nova.network.neutron [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updated VIF entry in instance network info cache for port ca7a22d5-2b52-42fe-a439-1ae3808aa672. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1154.154852] env[62066]: DEBUG nova.network.neutron [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating instance_info_cache with network_info: [{"id": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "address": "fa:16:3e:ce:29:d2", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca7a22d5-2b", "ovs_interfaceid": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.165556] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5234377f-30cd-32fa-5d53-63be7461af0b, 'name': SearchDatastore_Task, 'duration_secs': 0.009481} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.166360] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.166589] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.166819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.166968] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.167194] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.167671] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-809cd8a1-6de0-4dda-b042-2df3d5b9e1c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.175311] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.175463] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1154.176157] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78fe47da-46e1-4a17-b0eb-02abdae98a0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.181087] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1154.181087] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52a71ef0-91fe-4aae-ff01-12200e753d17" [ 1154.181087] env[62066]: _type = "Task" [ 1154.181087] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.188465] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a71ef0-91fe-4aae-ff01-12200e753d17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.660442] env[62066]: DEBUG oslo_concurrency.lockutils [req-ac430c2d-b83e-41f3-8756-0c4bf5eef114 req-f09542c3-d45d-4aa6-8521-db250201670c service nova] Releasing lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.691434] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52a71ef0-91fe-4aae-ff01-12200e753d17, 'name': SearchDatastore_Task, 'duration_secs': 0.007561} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.692180] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7af7277-a0b8-4cb5-ab58-d8f2909b6a80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.697276] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1154.697276] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52384e22-d3b3-5dca-9642-f273f21bfc33" [ 1154.697276] env[62066]: _type = "Task" [ 1154.697276] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.704316] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52384e22-d3b3-5dca-9642-f273f21bfc33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.207867] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52384e22-d3b3-5dca-9642-f273f21bfc33, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.208165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.208497] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 209fc51b-a036-4deb-9242-258f9c3e955c/209fc51b-a036-4deb-9242-258f9c3e955c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1155.208768] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a6dcc9f-75f1-41a6-90c1-c7143c84aa3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.214772] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1155.214772] env[62066]: value = "task-1156853" [ 1155.214772] env[62066]: _type = "Task" [ 1155.214772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.222472] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.725079] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455354} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.725435] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 209fc51b-a036-4deb-9242-258f9c3e955c/209fc51b-a036-4deb-9242-258f9c3e955c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1155.725607] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1155.725798] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43e22f1c-a95a-4bf4-916b-cbbd59f24baf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.731788] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1155.731788] env[62066]: value = "task-1156854" [ 1155.731788] env[62066]: _type = "Task" [ 1155.731788] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.739726] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156854, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.241659] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156854, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.241906] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.242657] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04dc008c-f31f-4e52-8a9d-18df1768fa5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.263840] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 209fc51b-a036-4deb-9242-258f9c3e955c/209fc51b-a036-4deb-9242-258f9c3e955c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.264075] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80277957-7f01-49a2-b6f6-e894c557c911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.284067] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1156.284067] env[62066]: value = "task-1156855" [ 1156.284067] env[62066]: _type = "Task" [ 1156.284067] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.292569] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.794443] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156855, 'name': ReconfigVM_Task, 'duration_secs': 0.312448} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.794832] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 209fc51b-a036-4deb-9242-258f9c3e955c/209fc51b-a036-4deb-9242-258f9c3e955c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.795498] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b09946a2-261d-46bc-97c9-49d3bb8b20ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.801802] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1156.801802] env[62066]: value = "task-1156856" [ 1156.801802] env[62066]: _type = "Task" [ 1156.801802] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.809407] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156856, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.312339] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156856, 'name': Rename_Task, 'duration_secs': 0.134137} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.312682] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.312983] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53d3dd96-a6fe-47a9-972c-8d1fdd4c1d81 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.318745] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1157.318745] env[62066]: value = "task-1156857" [ 1157.318745] env[62066]: _type = "Task" [ 1157.318745] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.326411] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156857, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.828931] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156857, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.329117] env[62066]: DEBUG oslo_vmware.api [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156857, 'name': PowerOnVM_Task, 'duration_secs': 0.525317} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.329468] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1158.329593] env[62066]: INFO nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1158.329840] env[62066]: DEBUG nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1158.330549] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ed857f-4321-4432-aebd-50428fb18a01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.846990] env[62066]: INFO nova.compute.manager [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Took 11.84 seconds to build instance. [ 1159.306217] env[62066]: DEBUG nova.compute.manager [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Received event network-changed-ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1159.306367] env[62066]: DEBUG nova.compute.manager [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Refreshing instance network info cache due to event network-changed-ca7a22d5-2b52-42fe-a439-1ae3808aa672. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1159.306626] env[62066]: DEBUG oslo_concurrency.lockutils [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] Acquiring lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.306746] env[62066]: DEBUG oslo_concurrency.lockutils [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] Acquired lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.306945] env[62066]: DEBUG nova.network.neutron [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Refreshing network info cache for port ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1159.348757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea2fa64d-bab0-4f43-a0de-00103cd542b0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.345s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.019584] env[62066]: DEBUG nova.network.neutron [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updated VIF entry in instance network info cache for port ca7a22d5-2b52-42fe-a439-1ae3808aa672. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1160.019895] env[62066]: DEBUG nova.network.neutron [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating instance_info_cache with network_info: [{"id": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "address": "fa:16:3e:ce:29:d2", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca7a22d5-2b", "ovs_interfaceid": "ca7a22d5-2b52-42fe-a439-1ae3808aa672", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.522053] env[62066]: DEBUG oslo_concurrency.lockutils [req-59d9b6e9-13f6-4252-8ce5-976338dfec51 req-fabb2d68-0bd5-4018-a71c-6ab46d9c1450 service nova] Releasing lock "refresh_cache-209fc51b-a036-4deb-9242-258f9c3e955c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.929983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.930332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.930555] env[62066]: INFO nova.compute.manager [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Rebooting instance [ 1182.448537] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.448739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.448925] env[62066]: DEBUG nova.network.neutron [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1183.146438] env[62066]: DEBUG nova.network.neutron [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.649443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.652911] env[62066]: DEBUG nova.compute.manager [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1183.653783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a98238-dba2-4c24-b11c-a94f61a2b06e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.669762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12f9dfd-e9ac-442b-b15f-8b2e0eb27f9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.677130] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Doing hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1184.677421] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-f21f4b65-5a32-46dc-b177-56ddea849635 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.683349] env[62066]: DEBUG oslo_vmware.api [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1184.683349] env[62066]: value = "task-1156858" [ 1184.683349] env[62066]: _type = "Task" [ 1184.683349] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.690471] env[62066]: DEBUG oslo_vmware.api [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156858, 'name': ResetVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.193347] env[62066]: DEBUG oslo_vmware.api [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156858, 'name': ResetVM_Task, 'duration_secs': 0.080024} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.193609] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Did hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1185.193752] env[62066]: DEBUG nova.compute.manager [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.194514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909b0833-81c8-440a-96f2-d0b381e81736 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.705925] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c6095eea-d063-4c96-abea-4a20940e08f5 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.775s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.698707] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.201680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.202175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.202175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.202320] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1187.203187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb613199-9bf1-46dd-b175-295d69a213be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.211250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faf8f29-09b2-455d-9ff9-78ab4561ca8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.224880] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7a3167-109b-4afd-a8ce-0b5ec2a7d0ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.231235] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56c72ad-53b3-46b3-a97c-fbe005b71648 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.259673] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181155MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1187.259887] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.259999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.284453] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01135457-b646-4c30-a730-1a18468ce58a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1188.284724] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 209fc51b-a036-4deb-9242-258f9c3e955c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1188.284820] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1188.284948] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1188.320602] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6cca1b-9432-4c26-a527-2497d1186ba6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.328762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e418e3-233b-4ab2-85a2-ed2ca6f3772e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.358749] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7894c8-a8db-44c5-adcb-b003dd0ab6ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.365183] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f989634-349a-4923-8b3d-e30eb95137c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.377670] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.880386] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.385517] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1189.386069] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.382069] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.382069] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.886157] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.886342] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1195.919904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.920126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.920216] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1196.556873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.557273] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.060592] env[62066]: DEBUG nova.compute.utils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1197.140880] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.563688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.643617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.643824] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1197.644045] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.644216] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.644442] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.644613] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.644759] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.644906] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.645049] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1198.625810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.626178] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.626322] env[62066]: INFO nova.compute.manager [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Attaching volume c3d44561-67b9-4e2c-b171-4b588f277fc4 to /dev/sdb [ 1198.656930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cbef8a-b6b2-4db6-81b4-0f97c801b929 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.664336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6580fbb4-17e4-4adc-a1d7-8d0d1b1a8354 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.677279] env[62066]: DEBUG nova.virt.block_device [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating existing volume attachment record: 4e31d37c-eeba-451b-9961-9739128c468b {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1203.224522] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1203.224777] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251799', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'name': 'volume-c3d44561-67b9-4e2c-b171-4b588f277fc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '209fc51b-a036-4deb-9242-258f9c3e955c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'serial': 'c3d44561-67b9-4e2c-b171-4b588f277fc4'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1203.225683] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7512fc-5154-4305-bb64-2d694445f4f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.242167] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe078b70-4828-4f53-8937-047a93113bae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.267491] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] volume-c3d44561-67b9-4e2c-b171-4b588f277fc4/volume-c3d44561-67b9-4e2c-b171-4b588f277fc4.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1203.267747] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44e6b3b7-1fdd-42fb-a642-9a05306376bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.285300] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1203.285300] env[62066]: value = "task-1156861" [ 1203.285300] env[62066]: _type = "Task" [ 1203.285300] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.294549] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156861, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.795106] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156861, 'name': ReconfigVM_Task, 'duration_secs': 0.327986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.795349] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfigured VM instance instance-00000072 to attach disk [datastore2] volume-c3d44561-67b9-4e2c-b171-4b588f277fc4/volume-c3d44561-67b9-4e2c-b171-4b588f277fc4.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1203.800835] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-035eb547-99bd-4d34-bc8b-6be1a032a0b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.815645] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1203.815645] env[62066]: value = "task-1156862" [ 1203.815645] env[62066]: _type = "Task" [ 1203.815645] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.826348] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156862, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.325915] env[62066]: DEBUG oslo_vmware.api [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156862, 'name': ReconfigVM_Task, 'duration_secs': 0.134727} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.326335] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251799', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'name': 'volume-c3d44561-67b9-4e2c-b171-4b588f277fc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '209fc51b-a036-4deb-9242-258f9c3e955c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'serial': 'c3d44561-67b9-4e2c-b171-4b588f277fc4'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1205.360870] env[62066]: DEBUG nova.objects.instance [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid 209fc51b-a036-4deb-9242-258f9c3e955c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.866760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e1e99ae-1c52-4589-832e-4c7ac535b2a7 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.072455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.072723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.575236] env[62066]: INFO nova.compute.manager [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Detaching volume c3d44561-67b9-4e2c-b171-4b588f277fc4 [ 1206.604637] env[62066]: INFO nova.virt.block_device [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Attempting to driver detach volume c3d44561-67b9-4e2c-b171-4b588f277fc4 from mountpoint /dev/sdb [ 1206.604881] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1206.605081] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251799', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'name': 'volume-c3d44561-67b9-4e2c-b171-4b588f277fc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '209fc51b-a036-4deb-9242-258f9c3e955c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'serial': 'c3d44561-67b9-4e2c-b171-4b588f277fc4'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1206.605941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97aced19-b421-4041-9e0c-d2889d388589 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.627266] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d653a2-f85a-4347-9331-036439b66046 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.633704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f944eb5d-a5b9-43ce-9f2d-c4b96857e38c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.653919] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e40c849-4d66-47ec-affa-5ee0f3f925d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.667607] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] The volume has not been displaced from its original location: [datastore2] volume-c3d44561-67b9-4e2c-b171-4b588f277fc4/volume-c3d44561-67b9-4e2c-b171-4b588f277fc4.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1206.672746] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1206.672991] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59b13b91-4f46-406d-8541-ce4aa37783a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.689502] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1206.689502] env[62066]: value = "task-1156863" [ 1206.689502] env[62066]: _type = "Task" [ 1206.689502] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.696574] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156863, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.199250] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156863, 'name': ReconfigVM_Task, 'duration_secs': 0.22218} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.199525] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.204011] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0782e8a-d63b-4018-b989-268a8983775c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.217489] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1207.217489] env[62066]: value = "task-1156864" [ 1207.217489] env[62066]: _type = "Task" [ 1207.217489] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.224669] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156864, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.727425] env[62066]: DEBUG oslo_vmware.api [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156864, 'name': ReconfigVM_Task, 'duration_secs': 0.118184} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.727746] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251799', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'name': 'volume-c3d44561-67b9-4e2c-b171-4b588f277fc4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '209fc51b-a036-4deb-9242-258f9c3e955c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c3d44561-67b9-4e2c-b171-4b588f277fc4', 'serial': 'c3d44561-67b9-4e2c-b171-4b588f277fc4'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1208.267872] env[62066]: DEBUG nova.objects.instance [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid 209fc51b-a036-4deb-9242-258f9c3e955c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.275904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64163ed6-cbe5-4854-a356-f4727d4e4b16 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.203s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.303675] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.304070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.304269] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.304483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.304669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.306973] env[62066]: INFO nova.compute.manager [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Terminating instance [ 1210.308817] env[62066]: DEBUG nova.compute.manager [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1210.309047] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1210.309996] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b939b4bf-9191-4d8e-89c7-5a569ed74206 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.319525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1210.319768] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8db24b17-7c0d-4a05-bf8b-11b03d59e738 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.325923] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1210.325923] env[62066]: value = "task-1156865" [ 1210.325923] env[62066]: _type = "Task" [ 1210.325923] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.335047] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.836111] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156865, 'name': PowerOffVM_Task, 'duration_secs': 0.178396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.836425] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1210.836620] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1210.836870] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81cf1182-3b0c-42d4-97be-8435f427e186 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.895289] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1210.895528] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1210.895715] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleting the datastore file [datastore2] 209fc51b-a036-4deb-9242-258f9c3e955c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.895982] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b423ad10-a45a-4e34-a837-a8e2ae49b249 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.902727] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1210.902727] env[62066]: value = "task-1156867" [ 1210.902727] env[62066]: _type = "Task" [ 1210.902727] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.909978] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.412728] env[62066]: DEBUG oslo_vmware.api [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145063} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.413144] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1211.413183] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1211.413361] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1211.413542] env[62066]: INFO nova.compute.manager [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1211.413782] env[62066]: DEBUG oslo.service.loopingcall [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1211.413994] env[62066]: DEBUG nova.compute.manager [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1211.414092] env[62066]: DEBUG nova.network.neutron [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1211.844846] env[62066]: DEBUG nova.compute.manager [req-3480bd81-04d7-441f-9ac6-bdc1fb9bc397 req-e9cab7c5-0b69-458b-810f-6597d5eb7310 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Received event network-vif-deleted-ca7a22d5-2b52-42fe-a439-1ae3808aa672 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1211.844977] env[62066]: INFO nova.compute.manager [req-3480bd81-04d7-441f-9ac6-bdc1fb9bc397 req-e9cab7c5-0b69-458b-810f-6597d5eb7310 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Neutron deleted interface ca7a22d5-2b52-42fe-a439-1ae3808aa672; detaching it from the instance and deleting it from the info cache [ 1211.845182] env[62066]: DEBUG nova.network.neutron [req-3480bd81-04d7-441f-9ac6-bdc1fb9bc397 req-e9cab7c5-0b69-458b-810f-6597d5eb7310 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.317389] env[62066]: DEBUG nova.network.neutron [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.348702] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10a772d9-975f-419e-997e-ec76d49781a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.360053] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93654874-1b50-415c-a34e-883f13c3b7b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.383451] env[62066]: DEBUG nova.compute.manager [req-3480bd81-04d7-441f-9ac6-bdc1fb9bc397 req-e9cab7c5-0b69-458b-810f-6597d5eb7310 service nova] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Detach interface failed, port_id=ca7a22d5-2b52-42fe-a439-1ae3808aa672, reason: Instance 209fc51b-a036-4deb-9242-258f9c3e955c could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1212.820353] env[62066]: INFO nova.compute.manager [-] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Took 1.41 seconds to deallocate network for instance. [ 1213.326928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.327243] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.327471] env[62066]: DEBUG nova.objects.instance [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'resources' on Instance uuid 209fc51b-a036-4deb-9242-258f9c3e955c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.870471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ebf37b-40dc-4b6a-81e2-65d727c3496e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.877797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64094b14-c665-4370-9f67-1cf8f665ba5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.908090] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d387a652-dde4-4a8c-97e7-23b945cdc491 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.914882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babe1d58-96d9-4486-bd15-f2e88506cb53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.927510] env[62066]: DEBUG nova.compute.provider_tree [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.430985] env[62066]: DEBUG nova.scheduler.client.report [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1214.936450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.957612] env[62066]: INFO nova.scheduler.client.report [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted allocations for instance 209fc51b-a036-4deb-9242-258f9c3e955c [ 1215.465957] env[62066]: DEBUG oslo_concurrency.lockutils [None req-11fcafe0-7b80-4540-9660-355322c000e0 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "209fc51b-a036-4deb-9242-258f9c3e955c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.162s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.111752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.112103] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.614805] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1218.134653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.134931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.136485] env[62066]: INFO nova.compute.claims [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1219.180196] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3caf0580-1d54-4fcb-b51b-ce02d83a0705 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.188161] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17219b67-e615-44ce-964f-fdd5d14537ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.217454] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c7c08f-e120-4d06-8f62-7713908ae926 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.224067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dca4c3-6656-498d-9819-4c45e24348e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.236700] env[62066]: DEBUG nova.compute.provider_tree [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.739703] env[62066]: DEBUG nova.scheduler.client.report [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1220.245875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.246461] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1220.752235] env[62066]: DEBUG nova.compute.utils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1220.753662] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1220.754229] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1220.806596] env[62066]: DEBUG nova.policy [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20873514f0fd4c4496d0cb5b2e500650', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a2bdc9041034d43b33453c202bd6cb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1221.077745] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Successfully created port: c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1221.258387] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1222.270035] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1222.290035] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1222.290035] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1222.290035] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1222.290259] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1222.290259] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1222.290396] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1222.290618] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1222.290793] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1222.290968] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1222.291187] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1222.291314] env[62066]: DEBUG nova.virt.hardware [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1222.292194] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee892f1a-48ef-4ab1-b0a3-9bfb92b52b32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.300222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360306e0-4829-45bd-bff6-e520a793fa44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.463774] env[62066]: DEBUG nova.compute.manager [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Received event network-vif-plugged-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1222.464080] env[62066]: DEBUG oslo_concurrency.lockutils [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.464259] env[62066]: DEBUG oslo_concurrency.lockutils [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.464435] env[62066]: DEBUG oslo_concurrency.lockutils [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.464609] env[62066]: DEBUG nova.compute.manager [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] No waiting events found dispatching network-vif-plugged-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1222.464778] env[62066]: WARNING nova.compute.manager [req-026eec7c-3083-4059-aeb1-f951a82f6a2b req-1ed4ac7b-ea04-42cb-a994-b1785784fd78 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Received unexpected event network-vif-plugged-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 for instance with vm_state building and task_state spawning. [ 1222.888218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "063c6f80-d621-4d64-912d-732b480843d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.888456] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.028225] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Successfully updated port: c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1223.051761] env[62066]: DEBUG nova.compute.manager [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Received event network-changed-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1223.051761] env[62066]: DEBUG nova.compute.manager [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Refreshing instance network info cache due to event network-changed-c7958dfa-28b0-4f7b-bc13-fc19aefe2782. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1223.051761] env[62066]: DEBUG oslo_concurrency.lockutils [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] Acquiring lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1223.051888] env[62066]: DEBUG oslo_concurrency.lockutils [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] Acquired lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.051928] env[62066]: DEBUG nova.network.neutron [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Refreshing network info cache for port c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1223.390694] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1223.533580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1223.586848] env[62066]: DEBUG nova.network.neutron [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1223.660543] env[62066]: DEBUG nova.network.neutron [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.912857] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.913135] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.914539] env[62066]: INFO nova.compute.claims [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1224.163350] env[62066]: DEBUG oslo_concurrency.lockutils [req-44bd9c22-ec44-4bbf-9b18-b92ff037356b req-31d36028-7e95-412a-9d81-4a1c37f52996 service nova] Releasing lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.163740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.163905] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1224.692431] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1224.809687] env[62066]: DEBUG nova.network.neutron [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating instance_info_cache with network_info: [{"id": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "address": "fa:16:3e:cb:57:e7", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7958dfa-28", "ovs_interfaceid": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.970328] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fb2acd-45f2-4330-a071-f09a69abb6eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.978034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebaa739-b39f-4ae8-acfe-66b510dc3b34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.006674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa1d775-0d7a-4504-844b-2b2c48573d49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.013358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08045f0-3584-4a67-90e9-5f6e998df05d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.025844] env[62066]: DEBUG nova.compute.provider_tree [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.312669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1225.312985] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance network_info: |[{"id": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "address": "fa:16:3e:cb:57:e7", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7958dfa-28", "ovs_interfaceid": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1225.313463] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:57:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7bb10726-a946-47b9-b4b5-6916e3f14cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7958dfa-28b0-4f7b-bc13-fc19aefe2782', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1225.320816] env[62066]: DEBUG oslo.service.loopingcall [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1225.321031] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1225.321257] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cf99480-bc66-4edb-b6a9-91730b2908c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.340431] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1225.340431] env[62066]: value = "task-1156868" [ 1225.340431] env[62066]: _type = "Task" [ 1225.340431] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.351196] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156868, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.528675] env[62066]: DEBUG nova.scheduler.client.report [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1225.850400] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156868, 'name': CreateVM_Task, 'duration_secs': 0.278429} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.850762] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1225.851214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.851390] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.851724] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1225.851971] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eb8b581-016a-4f51-90a8-ef39a0ee2dda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.856588] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1225.856588] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52866de6-8d91-016e-f8e5-125d13fd4320" [ 1225.856588] env[62066]: _type = "Task" [ 1225.856588] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.864088] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52866de6-8d91-016e-f8e5-125d13fd4320, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.033766] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.034313] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1226.366864] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52866de6-8d91-016e-f8e5-125d13fd4320, 'name': SearchDatastore_Task, 'duration_secs': 0.008906} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.367198] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1226.367442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1226.367675] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.367827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.368018] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1226.368281] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c82d704d-d4a2-479b-8d75-d2373aa9bc01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.375825] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1226.376014] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1226.376686] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-218ca492-3e4d-4324-bc0a-cfeafcba65ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.381336] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1226.381336] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52488c5b-745c-164d-7c40-55fa47f26195" [ 1226.381336] env[62066]: _type = "Task" [ 1226.381336] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.388651] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52488c5b-745c-164d-7c40-55fa47f26195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.539334] env[62066]: DEBUG nova.compute.utils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1226.540759] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1226.540928] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1226.586918] env[62066]: DEBUG nova.policy [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75a2f304461487e87c6d710eb7f830b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30c8aab967844a249c72080364b29278', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1226.820513] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Successfully created port: 4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1226.891827] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52488c5b-745c-164d-7c40-55fa47f26195, 'name': SearchDatastore_Task, 'duration_secs': 0.008207} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.892630] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c2483c5-4955-41c5-83f6-f1d4adb11143 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.897929] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1226.897929] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad2b89-f675-9cb9-979f-79fa03001dfd" [ 1226.897929] env[62066]: _type = "Task" [ 1226.897929] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.905505] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad2b89-f675-9cb9-979f-79fa03001dfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.044429] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1227.410059] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ad2b89-f675-9cb9-979f-79fa03001dfd, 'name': SearchDatastore_Task, 'duration_secs': 0.009812} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.410333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.410585] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3eda50c7-0d3d-47d0-b450-05fb5c059b60/3eda50c7-0d3d-47d0-b450-05fb5c059b60.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1227.410844] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b33e0bb-4bf4-4f34-af74-2f965fcd3585 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.417190] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1227.417190] env[62066]: value = "task-1156869" [ 1227.417190] env[62066]: _type = "Task" [ 1227.417190] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.424883] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.927167] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481878} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.927508] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 3eda50c7-0d3d-47d0-b450-05fb5c059b60/3eda50c7-0d3d-47d0-b450-05fb5c059b60.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1227.927588] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1227.927832] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3959df29-f6ec-4054-8d50-dc4492533e35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.933264] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1227.933264] env[62066]: value = "task-1156870" [ 1227.933264] env[62066]: _type = "Task" [ 1227.933264] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.940146] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.053624] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1228.077801] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1228.078073] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1228.078239] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1228.078428] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1228.078883] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1228.078883] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1228.079010] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1228.079134] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1228.079309] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1228.079473] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1228.079647] env[62066]: DEBUG nova.virt.hardware [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1228.080572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94808797-5a32-4c8f-b196-0e8f4bfec314 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.089610] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3a44ff-8c21-4848-8ba8-2ec5cfb8b7fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.182939] env[62066]: DEBUG nova.compute.manager [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Received event network-vif-plugged-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1228.183189] env[62066]: DEBUG oslo_concurrency.lockutils [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] Acquiring lock "063c6f80-d621-4d64-912d-732b480843d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.183401] env[62066]: DEBUG oslo_concurrency.lockutils [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] Lock "063c6f80-d621-4d64-912d-732b480843d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.183575] env[62066]: DEBUG oslo_concurrency.lockutils [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] Lock "063c6f80-d621-4d64-912d-732b480843d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1228.183743] env[62066]: DEBUG nova.compute.manager [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] No waiting events found dispatching network-vif-plugged-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1228.183954] env[62066]: WARNING nova.compute.manager [req-d34a0041-3eee-46c5-bcee-0898f1f52adc req-73e53c48-9abe-4208-8ac3-e0aa13f6a1a2 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Received unexpected event network-vif-plugged-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b for instance with vm_state building and task_state spawning. [ 1228.265786] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Successfully updated port: 4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1228.443831] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067011} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.444084] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1228.444893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33b69de-c753-4b0a-adbf-5527ca732e2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.466302] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 3eda50c7-0d3d-47d0-b450-05fb5c059b60/3eda50c7-0d3d-47d0-b450-05fb5c059b60.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1228.466540] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e71b8beb-0ec7-421b-b9d2-0e8efaa28ffa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.484952] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1228.484952] env[62066]: value = "task-1156871" [ 1228.484952] env[62066]: _type = "Task" [ 1228.484952] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.492187] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156871, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.767314] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1228.767497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.767617] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1228.995988] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156871, 'name': ReconfigVM_Task, 'duration_secs': 0.32641} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.996296] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 3eda50c7-0d3d-47d0-b450-05fb5c059b60/3eda50c7-0d3d-47d0-b450-05fb5c059b60.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1228.996836] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46e7ca1b-8f3c-44f5-8201-41ecc55f92d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.003831] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1229.003831] env[62066]: value = "task-1156872" [ 1229.003831] env[62066]: _type = "Task" [ 1229.003831] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.010790] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156872, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.300246] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1229.418900] env[62066]: DEBUG nova.network.neutron [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Updating instance_info_cache with network_info: [{"id": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "address": "fa:16:3e:ec:7d:da", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de5244c-07", "ovs_interfaceid": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.513435] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156872, 'name': Rename_Task, 'duration_secs': 0.140899} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.513731] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1229.513965] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6bdf11d-6658-48b5-b713-0af407f9e2a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.520041] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1229.520041] env[62066]: value = "task-1156873" [ 1229.520041] env[62066]: _type = "Task" [ 1229.520041] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.526770] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.922055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1229.922366] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance network_info: |[{"id": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "address": "fa:16:3e:ec:7d:da", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de5244c-07", "ovs_interfaceid": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1229.922830] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:7d:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4de5244c-07d6-4da8-ac4d-3682ac5f2b0b', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1229.930467] env[62066]: DEBUG oslo.service.loopingcall [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1229.930694] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1229.930927] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc192f4d-0dc9-4ef3-b750-567a804e2668 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.950638] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1229.950638] env[62066]: value = "task-1156874" [ 1229.950638] env[62066]: _type = "Task" [ 1229.950638] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.958176] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156874, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.029846] env[62066]: DEBUG oslo_vmware.api [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156873, 'name': PowerOnVM_Task, 'duration_secs': 0.441496} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.030222] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1230.030435] env[62066]: INFO nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Took 7.76 seconds to spawn the instance on the hypervisor. [ 1230.030669] env[62066]: DEBUG nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1230.031465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec572888-31f3-4e58-b122-27b3f65968cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.207987] env[62066]: DEBUG nova.compute.manager [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Received event network-changed-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1230.208228] env[62066]: DEBUG nova.compute.manager [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Refreshing instance network info cache due to event network-changed-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1230.208454] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] Acquiring lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.208601] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] Acquired lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.208795] env[62066]: DEBUG nova.network.neutron [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Refreshing network info cache for port 4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1230.460880] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156874, 'name': CreateVM_Task, 'duration_secs': 0.303614} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.461084] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1230.461756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.461931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.462278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1230.462531] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd330377-1790-48fc-ac54-f20422c2346e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.466985] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1230.466985] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52612cf3-5f5d-a51a-ea91-d91a24e12795" [ 1230.466985] env[62066]: _type = "Task" [ 1230.466985] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.474287] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52612cf3-5f5d-a51a-ea91-d91a24e12795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.548436] env[62066]: INFO nova.compute.manager [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Took 12.43 seconds to build instance. [ 1230.977843] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52612cf3-5f5d-a51a-ea91-d91a24e12795, 'name': SearchDatastore_Task, 'duration_secs': 0.008401} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.978305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.978430] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1230.978564] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.978709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.978918] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1230.979312] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43036242-b0f2-422a-9837-6e24d58e7a12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.990172] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1230.990466] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1230.994193] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c196674-d5c4-4ba6-8b23-dbf8a376c53d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.003319] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1231.003319] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5259bace-cf69-4ed7-b768-17db1280a704" [ 1231.003319] env[62066]: _type = "Task" [ 1231.003319] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.004505] env[62066]: DEBUG nova.network.neutron [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Updated VIF entry in instance network info cache for port 4de5244c-07d6-4da8-ac4d-3682ac5f2b0b. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1231.005212] env[62066]: DEBUG nova.network.neutron [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Updating instance_info_cache with network_info: [{"id": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "address": "fa:16:3e:ec:7d:da", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de5244c-07", "ovs_interfaceid": "4de5244c-07d6-4da8-ac4d-3682ac5f2b0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.021734] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5d8364e-7c86-4860-818e-c4f3c93a0513 req-e3b5cd4a-805f-47db-8b63-16b6540f6a43 service nova] Releasing lock "refresh_cache-063c6f80-d621-4d64-912d-732b480843d2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.028335] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5259bace-cf69-4ed7-b768-17db1280a704, 'name': SearchDatastore_Task, 'duration_secs': 0.008846} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.029137] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3d18935-f0d3-4c1b-931d-e9d9c62801a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.034226] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1231.034226] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52ef2bdb-1fb0-f709-c4b6-44f47d798962" [ 1231.034226] env[62066]: _type = "Task" [ 1231.034226] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.041617] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ef2bdb-1fb0-f709-c4b6-44f47d798962, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.050083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-51adc18f-93b8-4330-8537-83479a094182 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.362540] env[62066]: DEBUG nova.compute.manager [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Received event network-changed-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1231.362771] env[62066]: DEBUG nova.compute.manager [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Refreshing instance network info cache due to event network-changed-c7958dfa-28b0-4f7b-bc13-fc19aefe2782. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1231.362948] env[62066]: DEBUG oslo_concurrency.lockutils [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] Acquiring lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1231.363093] env[62066]: DEBUG oslo_concurrency.lockutils [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] Acquired lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.363261] env[62066]: DEBUG nova.network.neutron [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Refreshing network info cache for port c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1231.544618] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52ef2bdb-1fb0-f709-c4b6-44f47d798962, 'name': SearchDatastore_Task, 'duration_secs': 0.008256} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.544841] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.545116] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1231.545374] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ac38cea-eac6-4017-bdcd-7888bf48ce09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.552040] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1231.552040] env[62066]: value = "task-1156875" [ 1231.552040] env[62066]: _type = "Task" [ 1231.552040] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.559416] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.060949] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504496} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.061292] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1232.061450] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1232.061700] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fbb4631-027e-4af9-b6d6-672455e6c539 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.067855] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1232.067855] env[62066]: value = "task-1156876" [ 1232.067855] env[62066]: _type = "Task" [ 1232.067855] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.075416] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.111151] env[62066]: DEBUG nova.network.neutron [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updated VIF entry in instance network info cache for port c7958dfa-28b0-4f7b-bc13-fc19aefe2782. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1232.111547] env[62066]: DEBUG nova.network.neutron [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating instance_info_cache with network_info: [{"id": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "address": "fa:16:3e:cb:57:e7", "network": {"id": "b0ae91df-e7c8-4717-9dc6-3e372b293177", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-114999458-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1a2bdc9041034d43b33453c202bd6cb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7bb10726-a946-47b9-b4b5-6916e3f14cc5", "external-id": "nsx-vlan-transportzone-609", "segmentation_id": 609, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7958dfa-28", "ovs_interfaceid": "c7958dfa-28b0-4f7b-bc13-fc19aefe2782", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.577822] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.296071} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.578120] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1232.578907] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93253a00-3925-43f4-aedc-6dc57ea81d7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.599940] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1232.600177] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30ca5e5b-7ad9-4136-99d5-6fbc9f494745 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.614022] env[62066]: DEBUG oslo_concurrency.lockutils [req-499e854f-3850-406c-8e45-0d1bcb2919a3 req-e14f54dd-2d11-4d70-aa8f-e62146ec0e1b service nova] Releasing lock "refresh_cache-3eda50c7-0d3d-47d0-b450-05fb5c059b60" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.619869] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1232.619869] env[62066]: value = "task-1156877" [ 1232.619869] env[62066]: _type = "Task" [ 1232.619869] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.627226] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156877, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.130389] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156877, 'name': ReconfigVM_Task, 'duration_secs': 0.264442} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.130759] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1233.131228] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-593135be-e487-45ac-8b8d-3314138c5b6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.137267] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1233.137267] env[62066]: value = "task-1156878" [ 1233.137267] env[62066]: _type = "Task" [ 1233.137267] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.147678] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156878, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.648044] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156878, 'name': Rename_Task, 'duration_secs': 0.140018} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.648044] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1233.648044] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1529b77-e614-471d-b66c-5ef70d41e83d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.653344] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1233.653344] env[62066]: value = "task-1156879" [ 1233.653344] env[62066]: _type = "Task" [ 1233.653344] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.660305] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.163908] env[62066]: DEBUG oslo_vmware.api [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156879, 'name': PowerOnVM_Task, 'duration_secs': 0.421134} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.164299] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1234.164398] env[62066]: INFO nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Took 6.11 seconds to spawn the instance on the hypervisor. [ 1234.164565] env[62066]: DEBUG nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1234.165357] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020bfb4b-094a-4a78-ad6f-cd0f841bb948 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.680252] env[62066]: INFO nova.compute.manager [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Took 10.78 seconds to build instance. [ 1235.123559] env[62066]: INFO nova.compute.manager [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Rebuilding instance [ 1235.170442] env[62066]: DEBUG nova.compute.manager [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1235.171780] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3443b6-b26b-41cb-b854-347f314533fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.183871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b0b48652-0234-4f5a-8d06-3aa0ceb59968 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.295s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.686387] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1235.686714] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0db56002-9065-4cd5-b91c-dd97b5f8fcb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.694658] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1235.694658] env[62066]: value = "task-1156880" [ 1235.694658] env[62066]: _type = "Task" [ 1235.694658] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.703439] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.204497] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156880, 'name': PowerOffVM_Task, 'duration_secs': 0.195921} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.204875] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1236.205061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1236.205772] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9068df25-48a2-49cb-80d8-869e8deb8917 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.212604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1236.212848] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a0d297b-4e14-4c5c-81c0-c870cb7f4bb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.280696] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1236.281100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1236.281439] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleting the datastore file [datastore2] 063c6f80-d621-4d64-912d-732b480843d2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1236.281822] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-631323a2-83ab-4728-b58a-d45a1a51cfc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.290767] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1236.290767] env[62066]: value = "task-1156882" [ 1236.290767] env[62066]: _type = "Task" [ 1236.290767] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.302296] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.801867] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174056} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.803030] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1236.803030] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1236.803030] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1237.832937] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1237.833237] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1237.833389] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1237.833565] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1237.833715] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1237.833867] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1237.834094] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1237.834265] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1237.834437] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1237.834604] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1237.834780] env[62066]: DEBUG nova.virt.hardware [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1237.835639] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c210c006-e87f-4206-9e19-7cc189b87a87 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.843556] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804e4625-d5ec-4dd0-8063-665b7ad55b5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.857758] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:7d:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4de5244c-07d6-4da8-ac4d-3682ac5f2b0b', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1237.864921] env[62066]: DEBUG oslo.service.loopingcall [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1237.865161] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1237.865366] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e05705cb-2d70-4cb8-92f4-2539fa12e704 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.883594] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1237.883594] env[62066]: value = "task-1156883" [ 1237.883594] env[62066]: _type = "Task" [ 1237.883594] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.890912] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156883, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.392804] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156883, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.894321] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156883, 'name': CreateVM_Task, 'duration_secs': 0.960996} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.894666] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1238.895211] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.895393] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.895797] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1238.896074] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-464786aa-472a-4157-9ca3-b9f206bfcf16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.900543] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1238.900543] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5216f185-3047-d409-9428-1a8875846983" [ 1238.900543] env[62066]: _type = "Task" [ 1238.900543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.908041] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5216f185-3047-d409-9428-1a8875846983, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.410928] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5216f185-3047-d409-9428-1a8875846983, 'name': SearchDatastore_Task, 'duration_secs': 0.011654} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.411193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.411444] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1239.411684] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.411836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.412020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1239.412281] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbbc1368-aae1-434f-8950-6b4b10827587 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.420130] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1239.420295] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1239.420960] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3c0cfa6-62fe-4540-b65b-451cc84b7f18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.425562] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1239.425562] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5283430c-2b78-07ce-ea2e-a145715f959b" [ 1239.425562] env[62066]: _type = "Task" [ 1239.425562] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.432657] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5283430c-2b78-07ce-ea2e-a145715f959b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.935726] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5283430c-2b78-07ce-ea2e-a145715f959b, 'name': SearchDatastore_Task, 'duration_secs': 0.009068} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.936531] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3894e596-cd58-48f1-b994-7b656139053e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.941278] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1239.941278] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52675e8b-571c-d047-9230-86d37712cfea" [ 1239.941278] env[62066]: _type = "Task" [ 1239.941278] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.948546] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52675e8b-571c-d047-9230-86d37712cfea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.451523] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52675e8b-571c-d047-9230-86d37712cfea, 'name': SearchDatastore_Task, 'duration_secs': 0.009586} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.451748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1240.452020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1240.452273] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ed5f4fa-f548-4784-a7bf-bf6cbf66f848 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.460202] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1240.460202] env[62066]: value = "task-1156884" [ 1240.460202] env[62066]: _type = "Task" [ 1240.460202] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.467405] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.969769] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.406564} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.970136] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore1] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1240.970253] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1240.970496] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ea2186d-41f6-4244-af40-e8c081cbd472 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.977066] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1240.977066] env[62066]: value = "task-1156885" [ 1240.977066] env[62066]: _type = "Task" [ 1240.977066] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.984116] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156885, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.486718] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156885, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.487015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1241.487808] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879bea7b-0bc4-4acb-90f2-d8e41ca40d64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.509520] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1241.509761] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1bec3eb-c1ac-4b8b-bbea-e618a1abaf39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.528940] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1241.528940] env[62066]: value = "task-1156886" [ 1241.528940] env[62066]: _type = "Task" [ 1241.528940] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.536859] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156886, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.038833] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156886, 'name': ReconfigVM_Task, 'duration_secs': 0.261142} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.039206] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 063c6f80-d621-4d64-912d-732b480843d2/063c6f80-d621-4d64-912d-732b480843d2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1242.039809] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-131ded63-f969-451b-9762-71262f8801ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.046039] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1242.046039] env[62066]: value = "task-1156887" [ 1242.046039] env[62066]: _type = "Task" [ 1242.046039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.053551] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156887, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.556353] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156887, 'name': Rename_Task, 'duration_secs': 0.150377} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.556646] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1242.556886] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7acd18f2-f476-4ad7-9bf9-84c1460eb1b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.562385] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1242.562385] env[62066]: value = "task-1156888" [ 1242.562385] env[62066]: _type = "Task" [ 1242.562385] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.569393] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.072829] env[62066]: DEBUG oslo_vmware.api [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156888, 'name': PowerOnVM_Task, 'duration_secs': 0.426205} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.073233] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1243.073375] env[62066]: DEBUG nova.compute.manager [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1243.074095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4110f2-34b1-446f-8169-c926790906cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.591871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.592149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.592358] env[62066]: DEBUG nova.objects.instance [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1244.602184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ff969e7c-e3ed-4b3f-bc9e-1bea1f8cda5e tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.615381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "063c6f80-d621-4d64-912d-732b480843d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.615673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.615911] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "063c6f80-d621-4d64-912d-732b480843d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.616148] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.616365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.618787] env[62066]: INFO nova.compute.manager [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Terminating instance [ 1244.620573] env[62066]: DEBUG nova.compute.manager [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1244.620799] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1244.621738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3893b0e-f3ff-47b9-91bb-6c0fadb1e9d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.630597] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1244.630847] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-211817a1-da26-4cdb-8553-4c7293484977 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.636884] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1244.636884] env[62066]: value = "task-1156889" [ 1244.636884] env[62066]: _type = "Task" [ 1244.636884] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.644610] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.147195] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156889, 'name': PowerOffVM_Task, 'duration_secs': 0.180111} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.147473] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1245.147645] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1245.147888] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb21c273-14dc-45cd-a6b5-80a2e023f147 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.203215] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1245.203461] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1245.203645] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleting the datastore file [datastore1] 063c6f80-d621-4d64-912d-732b480843d2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1245.203931] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fa35f06-2d97-42c3-8042-e2518640fe68 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.210226] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1245.210226] env[62066]: value = "task-1156891" [ 1245.210226] env[62066]: _type = "Task" [ 1245.210226] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.217861] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.720409] env[62066]: DEBUG oslo_vmware.api [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154433} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.720903] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1245.720903] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1245.721063] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1245.721180] env[62066]: INFO nova.compute.manager [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1245.721437] env[62066]: DEBUG oslo.service.loopingcall [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1245.721639] env[62066]: DEBUG nova.compute.manager [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1245.721728] env[62066]: DEBUG nova.network.neutron [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1245.976589] env[62066]: DEBUG nova.compute.manager [req-acf53cbe-d563-4655-92dc-fdc8013e5570 req-ea6cf062-b3cc-45fe-8adf-1ff5281abef1 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Received event network-vif-deleted-4de5244c-07d6-4da8-ac4d-3682ac5f2b0b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1245.976909] env[62066]: INFO nova.compute.manager [req-acf53cbe-d563-4655-92dc-fdc8013e5570 req-ea6cf062-b3cc-45fe-8adf-1ff5281abef1 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Neutron deleted interface 4de5244c-07d6-4da8-ac4d-3682ac5f2b0b; detaching it from the instance and deleting it from the info cache [ 1245.976962] env[62066]: DEBUG nova.network.neutron [req-acf53cbe-d563-4655-92dc-fdc8013e5570 req-ea6cf062-b3cc-45fe-8adf-1ff5281abef1 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.450225] env[62066]: DEBUG nova.network.neutron [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.479953] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ee09a07-8d70-469c-8919-3aa6413c1f58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.489256] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e8d7c2-b729-4653-99a4-f1abc708c6c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.514620] env[62066]: DEBUG nova.compute.manager [req-acf53cbe-d563-4655-92dc-fdc8013e5570 req-ea6cf062-b3cc-45fe-8adf-1ff5281abef1 service nova] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Detach interface failed, port_id=4de5244c-07d6-4da8-ac4d-3682ac5f2b0b, reason: Instance 063c6f80-d621-4d64-912d-732b480843d2 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1246.698536] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.953116] env[62066]: INFO nova.compute.manager [-] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Took 1.23 seconds to deallocate network for instance. [ 1247.202037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.202312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.202431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.202583] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1247.203930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5266bba2-5ba7-4a14-8c99-5bbd76b0cc02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.212036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6b5fc0-eb90-43b4-94ad-624563aba08d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.225916] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da04927-bbc8-4111-9d97-16e96a373060 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.232337] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4f198e-9e81-4ec9-a4e1-b318b89870df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.260798] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181319MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1247.260935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.261139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.459511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.286975] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 01135457-b646-4c30-a730-1a18468ce58a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1248.287240] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 3eda50c7-0d3d-47d0-b450-05fb5c059b60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1248.287378] env[62066]: WARNING nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 063c6f80-d621-4d64-912d-732b480843d2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1248.287562] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1248.287738] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1248.332087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b85aac-9487-4c81-9da0-14cadc5fcfef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.338968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc59dfd9-9067-432d-98fc-6aae2c410a18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.369220] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e168c87-1c91-4978-8b54-2dcb76df2797 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.376112] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cd9a0a-df0d-4328-858e-d120fb579d7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.388630] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.891395] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1249.396840] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1249.397256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.136s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.397375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.938s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.397556] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.417285] env[62066]: INFO nova.scheduler.client.report [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted allocations for instance 063c6f80-d621-4d64-912d-732b480843d2 [ 1249.925113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f980c92-df2b-4343-a30f-21183ab292e7 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "063c6f80-d621-4d64-912d-732b480843d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.309s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.164888] env[62066]: DEBUG nova.compute.manager [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1251.684185] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.684488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.189684] env[62066]: INFO nova.compute.claims [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1252.696158] env[62066]: INFO nova.compute.resource_tracker [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating resource usage from migration b37a6575-80cb-41af-a7ac-d8596615ffd4 [ 1252.740161] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3875086c-a973-4e5b-b251-fd21e56e2144 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.747772] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8657272c-20b7-4c33-b152-148243cb896d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.777188] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29776784-a662-4c1a-8fc6-bb573b7cd281 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.783840] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f9300c-2d75-45cc-a4d0-b3ab36169e59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.796499] env[62066]: DEBUG nova.compute.provider_tree [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1253.299186] env[62066]: DEBUG nova.scheduler.client.report [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1253.805822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.121s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.806075] env[62066]: INFO nova.compute.manager [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Migrating [ 1254.321673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.321673] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.321673] env[62066]: DEBUG nova.network.neutron [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1255.030900] env[62066]: DEBUG nova.network.neutron [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.533957] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.394999] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1256.395264] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1256.395405] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1256.395523] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1256.899578] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.899954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.899954] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1256.900038] env[62066]: DEBUG nova.objects.instance [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lazy-loading 'info_cache' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.047653] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666b4e69-80e4-49cd-8550-9f04135539c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.067624] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1257.573144] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1257.573462] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4468789-0236-4b12-96dc-df34903cbaef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.581155] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1257.581155] env[62066]: value = "task-1156892" [ 1257.581155] env[62066]: _type = "Task" [ 1257.581155] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.589050] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156892, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.090126] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156892, 'name': PowerOffVM_Task, 'duration_secs': 0.178291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.090526] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1258.090630] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1258.596160] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1258.596425] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1258.596589] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1258.596774] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1258.596924] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1258.597088] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1258.597326] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1258.597488] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1258.597636] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1258.597800] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1258.597975] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1258.606291] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.607628] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8be0e46f-f9d3-4e94-9529-1f2da13b9d2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.624800] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1258.624800] env[62066]: value = "task-1156893" [ 1258.624800] env[62066]: _type = "Task" [ 1258.624800] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.632968] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156893, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.119973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.121905] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.121905] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1259.134725] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156893, 'name': ReconfigVM_Task, 'duration_secs': 0.171148} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.135423] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1259.641863] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1259.642155] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1259.642304] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1259.642489] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1259.642640] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1259.642792] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1259.642997] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1259.643176] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1259.643350] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1259.643518] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1259.643693] env[62066]: DEBUG nova.virt.hardware [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1259.648916] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1259.649226] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad369ccc-3292-4b6d-b732-8f64c6a61260 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.667443] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1259.667443] env[62066]: value = "task-1156894" [ 1259.667443] env[62066]: _type = "Task" [ 1259.667443] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.675082] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.177047] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156894, 'name': ReconfigVM_Task, 'duration_secs': 0.164051} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.177335] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1260.178077] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a527e7ae-9641-4a46-893f-8adb09cad868 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.198835] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.199068] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dabe12b0-deea-4618-888b-9f6c891ecf77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.216533] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1260.216533] env[62066]: value = "task-1156895" [ 1260.216533] env[62066]: _type = "Task" [ 1260.216533] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.223938] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.726685] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.227644] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156895, 'name': ReconfigVM_Task, 'duration_secs': 0.810566} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.227952] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 01135457-b646-4c30-a730-1a18468ce58a/01135457-b646-4c30-a730-1a18468ce58a.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.228271] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1261.736052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea4e169-c2dd-497d-bf9c-f41d78edf856 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.755198] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf8bd2f-83f2-4403-ba1f-c763bfb45c4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.772435] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1262.328211] env[62066]: DEBUG nova.network.neutron [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Port a30b3832-4a91-4ada-8984-d7cbc7d05969 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1263.348931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.348931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.349275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.393121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.393354] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.393496] env[62066]: DEBUG nova.network.neutron [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1265.107333] env[62066]: DEBUG nova.network.neutron [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.610456] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1266.134850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546d542e-cbaf-47b4-bf4e-857c0f89d3d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.153635] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef3df86-1beb-480e-8092-ac29a35c7467 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.160228] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1266.666594] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1266.666948] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d202c1d-ecef-4fdd-a3b3-570be5aafb9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.675772] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1266.675772] env[62066]: value = "task-1156896" [ 1266.675772] env[62066]: _type = "Task" [ 1266.675772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.683542] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.185742] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156896, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.685764] env[62066]: DEBUG oslo_vmware.api [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156896, 'name': PowerOnVM_Task, 'duration_secs': 0.787515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.686158] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1267.686239] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1d282f5-533c-4c07-b2ad-10581b8ff338 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance '01135457-b646-4c30-a730-1a18468ce58a' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1269.456170] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.456577] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.456631] env[62066]: DEBUG nova.compute.manager [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Going to confirm migration 8 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1269.834568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.834846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.004570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1270.004755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1270.004940] env[62066]: DEBUG nova.network.neutron [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1270.005154] env[62066]: DEBUG nova.objects.instance [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'info_cache' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.338078] env[62066]: DEBUG nova.compute.utils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1270.841517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.216069] env[62066]: DEBUG nova.network.neutron [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [{"id": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "address": "fa:16:3e:4d:09:66", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa30b3832-4a", "ovs_interfaceid": "a30b3832-4a91-4ada-8984-d7cbc7d05969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.718909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-01135457-b646-4c30-a730-1a18468ce58a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1271.719217] env[62066]: DEBUG nova.objects.instance [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'migration_context' on Instance uuid 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1271.901878] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.902257] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.902400] env[62066]: INFO nova.compute.manager [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Attaching volume 21f59fcb-0fc2-474b-b170-acce66ced6f2 to /dev/sdb [ 1271.931993] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3631f5a0-5cee-45cd-80be-4a23daf7b7f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.939617] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece874be-0d0a-4288-9fc6-56462d0491b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.953058] env[62066]: DEBUG nova.virt.block_device [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating existing volume attachment record: 2d6335d5-64ce-4bcd-ab4e-b1d4713c5ec4 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1272.222872] env[62066]: DEBUG nova.objects.base [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Object Instance<01135457-b646-4c30-a730-1a18468ce58a> lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1272.223613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261f02aa-50fe-4599-84a9-f149cdaa906b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.242694] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-098137a4-f3e4-4790-8b9b-75d4bf4bc84c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.247785] env[62066]: DEBUG oslo_vmware.api [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1272.247785] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8ef90-b18f-e652-9fc5-dfd0dc6f9cd7" [ 1272.247785] env[62066]: _type = "Task" [ 1272.247785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.255289] env[62066]: DEBUG oslo_vmware.api [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8ef90-b18f-e652-9fc5-dfd0dc6f9cd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.758281] env[62066]: DEBUG oslo_vmware.api [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52d8ef90-b18f-e652-9fc5-dfd0dc6f9cd7, 'name': SearchDatastore_Task, 'duration_secs': 0.006474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.758573] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.758812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.316482] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893dc9e1-cbf4-461c-afb2-d552e0f688db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.323676] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e92474-adf3-4ec4-bb32-b6b8326dc98b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.353288] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd309a9a-c222-4324-9fff-93ab8fd8044a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.359914] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b13ed0-a435-4fe6-a749-85993560832e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.372599] env[62066]: DEBUG nova.compute.provider_tree [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.875696] env[62066]: DEBUG nova.scheduler.client.report [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1274.886204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.127s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.439095] env[62066]: INFO nova.scheduler.client.report [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted allocation for migration b37a6575-80cb-41af-a7ac-d8596615ffd4 [ 1275.944283] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68db6676-4d3b-499c-80b0-29a90b589517 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.488s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.498336] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1276.498582] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251803', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'name': 'volume-21f59fcb-0fc2-474b-b170-acce66ced6f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3eda50c7-0d3d-47d0-b450-05fb5c059b60', 'attached_at': '', 'detached_at': '', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'serial': '21f59fcb-0fc2-474b-b170-acce66ced6f2'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1276.499462] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4ed681-2f7d-4c24-800c-2356ce89a233 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.515876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0fa3e1-a5e1-4e30-a316-83f67a20c4a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.541479] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] volume-21f59fcb-0fc2-474b-b170-acce66ced6f2/volume-21f59fcb-0fc2-474b-b170-acce66ced6f2.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1276.541720] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54f97bdb-e22a-41b6-96d3-612afe775432 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.559735] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1276.559735] env[62066]: value = "task-1156899" [ 1276.559735] env[62066]: _type = "Task" [ 1276.559735] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.567548] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156899, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.954419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.954716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.954920] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "01135457-b646-4c30-a730-1a18468ce58a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.955127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.955306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.957612] env[62066]: INFO nova.compute.manager [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Terminating instance [ 1276.959542] env[62066]: DEBUG nova.compute.manager [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1276.959770] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1276.960593] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc900405-2349-4421-88c0-8e3be5ff25d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.968636] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1276.968858] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c291ea3-a9b8-4840-a0c8-9622a68c76fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.975275] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1276.975275] env[62066]: value = "task-1156900" [ 1276.975275] env[62066]: _type = "Task" [ 1276.975275] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.982862] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.069547] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156899, 'name': ReconfigVM_Task, 'duration_secs': 0.30165} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.069868] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfigured VM instance instance-00000073 to attach disk [datastore2] volume-21f59fcb-0fc2-474b-b170-acce66ced6f2/volume-21f59fcb-0fc2-474b-b170-acce66ced6f2.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1277.074542] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d047af94-2614-4480-ae20-20f2d57efd9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.089551] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1277.089551] env[62066]: value = "task-1156901" [ 1277.089551] env[62066]: _type = "Task" [ 1277.089551] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.097476] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156901, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.485237] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156900, 'name': PowerOffVM_Task, 'duration_secs': 0.214718} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.485522] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1277.485658] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1277.485903] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32c5b214-a106-43f1-82ed-3917b7bcb7ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.550122] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1277.550363] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1277.550533] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleting the datastore file [datastore2] 01135457-b646-4c30-a730-1a18468ce58a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1277.550807] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a23d44a6-1792-4cb4-b3ac-a96c5f660368 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.557989] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1277.557989] env[62066]: value = "task-1156903" [ 1277.557989] env[62066]: _type = "Task" [ 1277.557989] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.565806] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.598571] env[62066]: DEBUG oslo_vmware.api [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156901, 'name': ReconfigVM_Task, 'duration_secs': 0.177034} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.598883] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251803', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'name': 'volume-21f59fcb-0fc2-474b-b170-acce66ced6f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3eda50c7-0d3d-47d0-b450-05fb5c059b60', 'attached_at': '', 'detached_at': '', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'serial': '21f59fcb-0fc2-474b-b170-acce66ced6f2'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1278.067491] env[62066]: DEBUG oslo_vmware.api [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144626} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.067860] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1278.067980] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1278.068154] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1278.068342] env[62066]: INFO nova.compute.manager [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1278.068585] env[62066]: DEBUG oslo.service.loopingcall [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1278.068782] env[62066]: DEBUG nova.compute.manager [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1278.068878] env[62066]: DEBUG nova.network.neutron [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1278.501605] env[62066]: DEBUG nova.compute.manager [req-b24e6dd1-38b6-4894-9bc9-8bbc7bc5a214 req-8a60670a-5a91-426a-b3bc-9a03032fcdeb service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Received event network-vif-deleted-a30b3832-4a91-4ada-8984-d7cbc7d05969 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1278.501898] env[62066]: INFO nova.compute.manager [req-b24e6dd1-38b6-4894-9bc9-8bbc7bc5a214 req-8a60670a-5a91-426a-b3bc-9a03032fcdeb service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Neutron deleted interface a30b3832-4a91-4ada-8984-d7cbc7d05969; detaching it from the instance and deleting it from the info cache [ 1278.502243] env[62066]: DEBUG nova.network.neutron [req-b24e6dd1-38b6-4894-9bc9-8bbc7bc5a214 req-8a60670a-5a91-426a-b3bc-9a03032fcdeb service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1278.637479] env[62066]: DEBUG nova.objects.instance [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid 3eda50c7-0d3d-47d0-b450-05fb5c059b60 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1278.976333] env[62066]: DEBUG nova.network.neutron [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1279.005343] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffd9d4d7-e5f6-41ca-b6f5-35e1e583cbb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.016233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937740ff-8c41-4129-8cfd-ac494f0f84df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.039995] env[62066]: DEBUG nova.compute.manager [req-b24e6dd1-38b6-4894-9bc9-8bbc7bc5a214 req-8a60670a-5a91-426a-b3bc-9a03032fcdeb service nova] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Detach interface failed, port_id=a30b3832-4a91-4ada-8984-d7cbc7d05969, reason: Instance 01135457-b646-4c30-a730-1a18468ce58a could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1279.143505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7780cc39-6d70-4617-8359-7392f5b7bcd8 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.321259] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.322215] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.479511] env[62066]: INFO nova.compute.manager [-] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Took 1.41 seconds to deallocate network for instance. [ 1279.826227] env[62066]: INFO nova.compute.manager [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Detaching volume 21f59fcb-0fc2-474b-b170-acce66ced6f2 [ 1279.855380] env[62066]: INFO nova.virt.block_device [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Attempting to driver detach volume 21f59fcb-0fc2-474b-b170-acce66ced6f2 from mountpoint /dev/sdb [ 1279.855626] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1279.855820] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251803', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'name': 'volume-21f59fcb-0fc2-474b-b170-acce66ced6f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3eda50c7-0d3d-47d0-b450-05fb5c059b60', 'attached_at': '', 'detached_at': '', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'serial': '21f59fcb-0fc2-474b-b170-acce66ced6f2'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1279.856687] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d848f974-328c-442e-b693-ea1643eef045 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.877967] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735e0cf4-f630-4874-93b5-ba44a083753e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.884734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c3c8bd-627a-4990-b46f-0a77831a379d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.904990] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be58d2f6-26a1-4044-b2f7-716450a648b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.919333] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] The volume has not been displaced from its original location: [datastore2] volume-21f59fcb-0fc2-474b-b170-acce66ced6f2/volume-21f59fcb-0fc2-474b-b170-acce66ced6f2.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1279.924393] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1279.924636] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e25c13cf-c84d-4675-9b57-f53105a565c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.941777] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1279.941777] env[62066]: value = "task-1156904" [ 1279.941777] env[62066]: _type = "Task" [ 1279.941777] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.949125] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156904, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.986350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.986623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.986819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.008765] env[62066]: INFO nova.scheduler.client.report [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted allocations for instance 01135457-b646-4c30-a730-1a18468ce58a [ 1280.452185] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156904, 'name': ReconfigVM_Task, 'duration_secs': 0.221509} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.452581] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1280.459972] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a16cbdf9-8475-4a19-bde4-16633e4b6f69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.481345] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1280.481345] env[62066]: value = "task-1156905" [ 1280.481345] env[62066]: _type = "Task" [ 1280.481345] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.492195] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.516295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4acb31b8-22d0-4c28-9fac-9dce34688e1d tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "01135457-b646-4c30-a730-1a18468ce58a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.562s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.991178] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156905, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.492809] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156905, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.822325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.822561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.992932] env[62066]: DEBUG oslo_vmware.api [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156905, 'name': ReconfigVM_Task, 'duration_secs': 1.144004} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.993261] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-251803', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'name': 'volume-21f59fcb-0fc2-474b-b170-acce66ced6f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3eda50c7-0d3d-47d0-b450-05fb5c059b60', 'attached_at': '', 'detached_at': '', 'volume_id': '21f59fcb-0fc2-474b-b170-acce66ced6f2', 'serial': '21f59fcb-0fc2-474b-b170-acce66ced6f2'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1282.325740] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1282.537349] env[62066]: DEBUG nova.objects.instance [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'flavor' on Instance uuid 3eda50c7-0d3d-47d0-b450-05fb5c059b60 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.845230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.845499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.846964] env[62066]: INFO nova.compute.claims [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1283.544020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3b284541-2c20-4519-b794-eb68a5b6dc40 tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.222s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.893664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb664e18-6d82-4c87-91ee-47c612b2fd23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.901513] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64fa3bc2-462b-4d3e-8872-4e047dc72df8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.931457] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8230947e-7084-49e4-9d5f-0d800b63918b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.938248] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7cfff9-e27d-410f-90cb-21384734c1c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.952471] env[62066]: DEBUG nova.compute.provider_tree [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1284.455838] env[62066]: DEBUG nova.scheduler.client.report [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.581353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.581755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.581843] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.581979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.582171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.584199] env[62066]: INFO nova.compute.manager [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Terminating instance [ 1284.585934] env[62066]: DEBUG nova.compute.manager [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1284.586152] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1284.586971] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b20d54-6dfe-46f2-927f-b2edbf4176a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.594353] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1284.594578] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04f853ca-6926-48b9-a865-1b3d3c59f75b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.600913] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1284.600913] env[62066]: value = "task-1156906" [ 1284.600913] env[62066]: _type = "Task" [ 1284.600913] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.608145] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.961154] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.115s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.961703] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1285.111117] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156906, 'name': PowerOffVM_Task, 'duration_secs': 0.210784} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.111388] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1285.111560] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1285.111806] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebaa9490-ea35-4c6c-8845-0b76bcf640d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.177940] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1285.178137] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1285.178318] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleting the datastore file [datastore2] 3eda50c7-0d3d-47d0-b450-05fb5c059b60 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1285.178566] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32c38e35-fcf1-466d-b0ec-45c74c976712 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.184319] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for the task: (returnval){ [ 1285.184319] env[62066]: value = "task-1156908" [ 1285.184319] env[62066]: _type = "Task" [ 1285.184319] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.191512] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.466924] env[62066]: DEBUG nova.compute.utils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1285.468547] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1285.468734] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1285.504469] env[62066]: DEBUG nova.policy [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75a2f304461487e87c6d710eb7f830b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30c8aab967844a249c72080364b29278', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1285.694090] env[62066]: DEBUG oslo_vmware.api [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Task: {'id': task-1156908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12409} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.694366] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1285.694541] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1285.694718] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1285.694889] env[62066]: INFO nova.compute.manager [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1285.695153] env[62066]: DEBUG oslo.service.loopingcall [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1285.695346] env[62066]: DEBUG nova.compute.manager [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1285.695438] env[62066]: DEBUG nova.network.neutron [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1285.763340] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Successfully created port: fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1285.971860] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1286.144331] env[62066]: DEBUG nova.compute.manager [req-d4afe656-2394-45f4-be35-85f0567276fd req-4596d8c9-c3bf-417b-a6fa-918fe0ee5175 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Received event network-vif-deleted-c7958dfa-28b0-4f7b-bc13-fc19aefe2782 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1286.144331] env[62066]: INFO nova.compute.manager [req-d4afe656-2394-45f4-be35-85f0567276fd req-4596d8c9-c3bf-417b-a6fa-918fe0ee5175 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Neutron deleted interface c7958dfa-28b0-4f7b-bc13-fc19aefe2782; detaching it from the instance and deleting it from the info cache [ 1286.144504] env[62066]: DEBUG nova.network.neutron [req-d4afe656-2394-45f4-be35-85f0567276fd req-4596d8c9-c3bf-417b-a6fa-918fe0ee5175 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.622016] env[62066]: DEBUG nova.network.neutron [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.647239] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39df7486-ab9a-42ea-80d5-7bcbb6535da4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.657296] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a829a78f-1876-4eb2-9965-dc7ae7230fee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.681228] env[62066]: DEBUG nova.compute.manager [req-d4afe656-2394-45f4-be35-85f0567276fd req-4596d8c9-c3bf-417b-a6fa-918fe0ee5175 service nova] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Detach interface failed, port_id=c7958dfa-28b0-4f7b-bc13-fc19aefe2782, reason: Instance 3eda50c7-0d3d-47d0-b450-05fb5c059b60 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1286.981677] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1287.009294] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1287.009573] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1287.009734] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1287.009969] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1287.010150] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1287.010303] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1287.010511] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1287.010671] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1287.010840] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1287.011016] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1287.011198] env[62066]: DEBUG nova.virt.hardware [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1287.012083] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9a8f70-aa3c-477e-a4ac-fbfeab71f93c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.020457] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee5a03e-c137-48a3-ab4f-b3e503b378be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.127295] env[62066]: INFO nova.compute.manager [-] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Took 1.43 seconds to deallocate network for instance. [ 1287.130652] env[62066]: DEBUG nova.compute.manager [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Received event network-vif-plugged-fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1287.130866] env[62066]: DEBUG oslo_concurrency.lockutils [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] Acquiring lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1287.131129] env[62066]: DEBUG oslo_concurrency.lockutils [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1287.131320] env[62066]: DEBUG oslo_concurrency.lockutils [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1287.131489] env[62066]: DEBUG nova.compute.manager [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] No waiting events found dispatching network-vif-plugged-fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1287.131654] env[62066]: WARNING nova.compute.manager [req-399742a5-c2ab-4245-bf48-74e026e1981f req-6861e456-6e2a-48c5-b1ed-c1417b778efb service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Received unexpected event network-vif-plugged-fbffa55a-c090-4dc3-9f90-0059f6143ff7 for instance with vm_state building and task_state spawning. [ 1287.635981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1287.636304] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1287.636532] env[62066]: DEBUG nova.objects.instance [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lazy-loading 'resources' on Instance uuid 3eda50c7-0d3d-47d0-b450-05fb5c059b60 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1287.697355] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Successfully updated port: fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1287.722379] env[62066]: DEBUG nova.compute.manager [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Received event network-changed-fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1287.722613] env[62066]: DEBUG nova.compute.manager [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Refreshing instance network info cache due to event network-changed-fbffa55a-c090-4dc3-9f90-0059f6143ff7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1287.722790] env[62066]: DEBUG oslo_concurrency.lockutils [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1287.722936] env[62066]: DEBUG oslo_concurrency.lockutils [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.723284] env[62066]: DEBUG nova.network.neutron [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Refreshing network info cache for port fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1288.182527] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba8a7e5-71ed-4058-88fc-a0ebf78fe359 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.189713] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4cbba6-6372-48c2-b964-a3fc74391f17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.218376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.219233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b32de8-3a7f-44d3-b079-2d62fbdc429e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.227618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57ffaa9-faef-4ccc-9956-787f0934f1c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.241630] env[62066]: DEBUG nova.compute.provider_tree [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1288.257748] env[62066]: DEBUG nova.network.neutron [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1288.324126] env[62066]: DEBUG nova.network.neutron [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.744465] env[62066]: DEBUG nova.scheduler.client.report [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1288.826389] env[62066]: DEBUG oslo_concurrency.lockutils [req-13766f4d-1192-424d-9928-83ac32ffba86 req-f5bd13a6-3dae-42e6-bc06-1e76cf106b85 service nova] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1288.826788] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.826962] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1289.249731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.273459] env[62066]: INFO nova.scheduler.client.report [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Deleted allocations for instance 3eda50c7-0d3d-47d0-b450-05fb5c059b60 [ 1289.356519] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1289.483626] env[62066]: DEBUG nova.network.neutron [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.783386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2bccbb6f-0283-4b70-b974-4c265c47d9ca tempest-AttachVolumeNegativeTest-1555166714 tempest-AttachVolumeNegativeTest-1555166714-project-member] Lock "3eda50c7-0d3d-47d0-b450-05fb5c059b60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.202s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1289.986274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.986613] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance network_info: |[{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1289.987090] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:76:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbffa55a-c090-4dc3-9f90-0059f6143ff7', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1289.994562] env[62066]: DEBUG oslo.service.loopingcall [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1289.994780] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1289.995018] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7ebd68b-bdaf-47b9-871f-dfbab2109c41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.013485] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1290.013485] env[62066]: value = "task-1156910" [ 1290.013485] env[62066]: _type = "Task" [ 1290.013485] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.025359] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156910, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.523829] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156910, 'name': CreateVM_Task, 'duration_secs': 0.351257} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.524253] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1290.524701] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1290.524880] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.525242] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1290.525493] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0623fcc8-b422-4011-a263-aa74624d8ae0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.529622] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1290.529622] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]523289fb-41aa-c4d3-3868-967433c13eca" [ 1290.529622] env[62066]: _type = "Task" [ 1290.529622] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.536817] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523289fb-41aa-c4d3-3868-967433c13eca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.039677] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]523289fb-41aa-c4d3-3868-967433c13eca, 'name': SearchDatastore_Task, 'duration_secs': 0.008575} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.039999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1291.040259] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1291.040495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1291.040647] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.040827] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1291.041121] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65e6cf25-df38-4afb-9ea9-2af57abcaa0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.049126] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1291.049311] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1291.049987] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5b2576c-7253-4bb1-9183-8932ffe05287 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.054590] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1291.054590] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5277ef99-7d67-a330-83c0-3c3843601aa9" [ 1291.054590] env[62066]: _type = "Task" [ 1291.054590] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.061654] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5277ef99-7d67-a330-83c0-3c3843601aa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.564798] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5277ef99-7d67-a330-83c0-3c3843601aa9, 'name': SearchDatastore_Task, 'duration_secs': 0.007631} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.565539] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b6c9890-c19d-43a0-b99f-b72b12a2acbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.570429] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1291.570429] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52f8a67d-312e-56da-740f-b9f93976a2e5" [ 1291.570429] env[62066]: _type = "Task" [ 1291.570429] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.577476] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f8a67d-312e-56da-740f-b9f93976a2e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.080629] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52f8a67d-312e-56da-740f-b9f93976a2e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008482} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.080903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1292.081202] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1292.081461] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-235f18cd-03b2-4cdd-a626-d95418576dd2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.087759] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1292.087759] env[62066]: value = "task-1156911" [ 1292.087759] env[62066]: _type = "Task" [ 1292.087759] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.095318] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.598162] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445976} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.598458] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1292.598617] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1292.598902] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34d62ec0-2271-4677-a627-f367e8f31a2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.604651] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1292.604651] env[62066]: value = "task-1156912" [ 1292.604651] env[62066]: _type = "Task" [ 1292.604651] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.611816] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.114416] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059665} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.114642] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1293.115519] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3644b2b3-4781-4ca0-8cc9-e887c339d7cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.137738] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1293.137984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe612b1c-4b70-4a05-90a0-3fa136143c56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.158076] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1293.158076] env[62066]: value = "task-1156914" [ 1293.158076] env[62066]: _type = "Task" [ 1293.158076] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.165555] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156914, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.668198] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156914, 'name': ReconfigVM_Task, 'duration_secs': 0.282156} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.668583] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfigured VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1293.669126] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae9dcd20-50f0-454f-8eb9-eee3aa673f34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.675019] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1293.675019] env[62066]: value = "task-1156915" [ 1293.675019] env[62066]: _type = "Task" [ 1293.675019] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.682097] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156915, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.184190] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156915, 'name': Rename_Task, 'duration_secs': 0.135139} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.184461] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1294.184700] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6e685d6-9904-4228-9c20-1ae5b1db5cf2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.190613] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1294.190613] env[62066]: value = "task-1156916" [ 1294.190613] env[62066]: _type = "Task" [ 1294.190613] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.197483] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.700690] env[62066]: DEBUG oslo_vmware.api [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156916, 'name': PowerOnVM_Task, 'duration_secs': 0.417889} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.701083] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1294.701200] env[62066]: INFO nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1294.701392] env[62066]: DEBUG nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1294.702162] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d10a13d-b5c8-499b-afe6-09078f8a788f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.219957] env[62066]: INFO nova.compute.manager [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Took 12.39 seconds to build instance. [ 1295.722523] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dff58609-7534-43ce-871f-00ccaf8c024f tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.900s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.039786] env[62066]: DEBUG nova.compute.manager [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Received event network-changed-fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1296.039949] env[62066]: DEBUG nova.compute.manager [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Refreshing instance network info cache due to event network-changed-fbffa55a-c090-4dc3-9f90-0059f6143ff7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1296.040933] env[62066]: DEBUG oslo_concurrency.lockutils [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1296.040933] env[62066]: DEBUG oslo_concurrency.lockutils [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1296.040933] env[62066]: DEBUG nova.network.neutron [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Refreshing network info cache for port fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1296.747673] env[62066]: DEBUG nova.network.neutron [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updated VIF entry in instance network info cache for port fbffa55a-c090-4dc3-9f90-0059f6143ff7. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1296.748119] env[62066]: DEBUG nova.network.neutron [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.250882] env[62066]: DEBUG oslo_concurrency.lockutils [req-71d90f30-6b48-4227-b514-6c5a55ed5e60 req-b001016f-8ae7-41c9-8597-a347df5914c8 service nova] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1308.699369] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1309.202160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.202394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1309.202566] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.202745] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1309.203671] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37b51b7-5880-4f8f-b576-e1503e8bb201 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.211712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e779c84b-e90f-4b15-bff9-e15b7225c375 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.225219] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91417ef-653c-4fab-b82b-f4b30106f872 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.231406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9141f8-6321-4065-8ccb-2d97981cb914 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.259834] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181282MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1309.259965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1309.260173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.285153] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1310.285454] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1310.285509] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1310.310218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af685c6c-bdb2-4850-9780-5a1851c3f20a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.317852] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63897e2-2d4c-49d8-a797-d6acb73fdabe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.348202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff5a01d-0972-4357-bab3-ff3feff32f76 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.355318] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ac217e-9f32-4da6-b9ab-b730a80e97a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.368359] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.871889] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1311.377345] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1311.377736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.117s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1317.372520] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.372948] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.878086] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.878086] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1318.380808] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1318.381211] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381260] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381373] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381521] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381662] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381806] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.381934] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1333.372058] env[62066]: DEBUG nova.compute.manager [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1334.121646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1334.121930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1334.627068] env[62066]: INFO nova.compute.claims [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1335.132920] env[62066]: INFO nova.compute.resource_tracker [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating resource usage from migration e72836b4-d9c3-4047-bf65-34021f25a467 [ 1335.169877] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8447a3-9b8d-4260-ac65-ade0cd92909b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.178065] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c551da93-a571-41bc-8e20-a5c5f79227ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.206683] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551f7ca0-3acd-4eda-a320-173248772a98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.213218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0311fbba-0d97-4609-83fb-83647a3c13a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.225607] env[62066]: DEBUG nova.compute.provider_tree [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1335.729519] env[62066]: DEBUG nova.scheduler.client.report [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1336.236775] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.115s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1336.236997] env[62066]: INFO nova.compute.manager [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Migrating [ 1336.752188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1336.752536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.752591] env[62066]: DEBUG nova.network.neutron [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1337.462522] env[62066]: DEBUG nova.network.neutron [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.964975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1339.480400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b063aa7c-2158-40b8-9233-8f095cff121f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.499167] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1340.004801] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1340.005169] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5009e9af-6101-44e6-b28c-a5f1474b737c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.012804] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1340.012804] env[62066]: value = "task-1156918" [ 1340.012804] env[62066]: _type = "Task" [ 1340.012804] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.020635] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156918, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.523770] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156918, 'name': PowerOffVM_Task, 'duration_secs': 0.182073} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.524204] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1340.524304] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1341.031336] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1341.031336] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1341.031336] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1341.031548] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1341.031611] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1341.031759] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1341.031962] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1341.032137] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1341.032303] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1341.032463] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1341.032629] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1341.037604] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97770e71-b341-4d43-8de9-c6c9099887dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.053639] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1341.053639] env[62066]: value = "task-1156919" [ 1341.053639] env[62066]: _type = "Task" [ 1341.053639] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.061572] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156919, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.562996] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156919, 'name': ReconfigVM_Task, 'duration_secs': 0.14485} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.563405] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1342.069991] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1342.070255] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1342.070423] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1342.070612] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1342.070758] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1342.070908] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1342.071157] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1342.071347] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1342.071521] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1342.071685] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1342.071861] env[62066]: DEBUG nova.virt.hardware [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1342.077186] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1342.077477] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57d1c396-d4d4-42a6-863e-671f6f88bfb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.095473] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1342.095473] env[62066]: value = "task-1156920" [ 1342.095473] env[62066]: _type = "Task" [ 1342.095473] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.103142] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156920, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.605459] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156920, 'name': ReconfigVM_Task, 'duration_secs': 0.160826} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.605865] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1342.606519] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e46ae5a-f60d-4178-b434-3c3f518430e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.627585] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1342.628128] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5f1a38d-7c55-4dee-9672-9f809c5106f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.645803] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1342.645803] env[62066]: value = "task-1156921" [ 1342.645803] env[62066]: _type = "Task" [ 1342.645803] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.652826] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.155309] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156921, 'name': ReconfigVM_Task, 'duration_secs': 0.244196} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.155597] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfigured VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1343.155860] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1343.663541] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95a240a-4fd1-487f-b1be-75f970f4aaab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.683558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac7910d-75aa-4f83-9291-ad4c9eb1c401 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.702236] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1344.244890] env[62066]: DEBUG nova.network.neutron [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Port fbffa55a-c090-4dc3-9f90-0059f6143ff7 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1345.267082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.267463] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.267581] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.300548] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1346.300827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1346.300933] env[62066]: DEBUG nova.network.neutron [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1347.004317] env[62066]: DEBUG nova.network.neutron [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1347.506954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1348.030807] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04202ca-5a01-4002-9487-3163c3d48cbe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.050406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ff1d64-f2af-41bb-b3d8-4f1fc8db825d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.057107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1348.563251] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1348.563631] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f418969-dd24-406f-a33a-76e6848eebc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.570195] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1348.570195] env[62066]: value = "task-1156922" [ 1348.570195] env[62066]: _type = "Task" [ 1348.570195] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.577754] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.079910] env[62066]: DEBUG oslo_vmware.api [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156922, 'name': PowerOnVM_Task, 'duration_secs': 0.385101} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.080246] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1349.080457] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3e7b49-5988-4db3-bbf4-094aa3fde92c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance '15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1351.756924] env[62066]: DEBUG nova.network.neutron [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Port fbffa55a-c090-4dc3-9f90-0059f6143ff7 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1351.757220] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1351.757357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1351.757529] env[62066]: DEBUG nova.network.neutron [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1352.505260] env[62066]: DEBUG nova.network.neutron [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1353.008275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1353.513527] env[62066]: DEBUG nova.compute.manager [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62066) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1353.513785] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.514035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.017661] env[62066]: DEBUG nova.objects.instance [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'migration_context' on Instance uuid 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1354.563041] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe64396-777a-4ff7-b354-a7d1ec96b6cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.570662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374d66db-57cb-41f0-90e1-81554d6a12e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.600764] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b631a7bb-ecc5-4ffb-872d-69524302fabc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.608323] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bbb364-9c26-40dd-829f-448fd7904c62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.621045] env[62066]: DEBUG nova.compute.provider_tree [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1355.124455] env[62066]: DEBUG nova.scheduler.client.report [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1356.136085] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.622s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.672414] env[62066]: INFO nova.compute.manager [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Swapping old allocation on dict_keys(['cd4c0e36-9c88-4f73-a93c-1ff383ed97c4']) held by migration e72836b4-d9c3-4047-bf65-34021f25a467 for instance [ 1357.692777] env[62066]: DEBUG nova.scheduler.client.report [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Overwriting current allocation {'allocations': {'cd4c0e36-9c88-4f73-a93c-1ff383ed97c4': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 163}}, 'project_id': '30c8aab967844a249c72080364b29278', 'user_id': 'd75a2f304461487e87c6d710eb7f830b', 'consumer_generation': 1} on consumer 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 {{(pid=62066) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1357.773546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.773762] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.773943] env[62066]: DEBUG nova.network.neutron [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1358.486454] env[62066]: DEBUG nova.network.neutron [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [{"id": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "address": "fa:16:3e:71:76:a7", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbffa55a-c0", "ovs_interfaceid": "fbffa55a-c090-4dc3-9f90-0059f6143ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.989919] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.990379] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1358.990706] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f84bbb2-119e-4bd3-8e60-995bdb7f4278 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.998697] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1358.998697] env[62066]: value = "task-1156923" [ 1358.998697] env[62066]: _type = "Task" [ 1358.998697] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.007174] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.508605] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156923, 'name': PowerOffVM_Task, 'duration_secs': 0.220197} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.508913] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1359.509606] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1359.509837] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1359.510009] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1359.510210] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1359.510362] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1359.510514] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1359.510723] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1359.510890] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1359.511084] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1359.511260] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1359.511466] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1359.516494] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2018282-b7d9-4d69-bb4c-5792b87e4d78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.532103] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1359.532103] env[62066]: value = "task-1156924" [ 1359.532103] env[62066]: _type = "Task" [ 1359.532103] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.539245] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.042201] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156924, 'name': ReconfigVM_Task, 'duration_secs': 0.122385} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.043063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47888f5b-f21a-49e2-b4a7-f99642d7297e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.061635] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1360.061890] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1360.062064] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1360.062261] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1360.062610] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1360.062610] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1360.062808] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1360.062943] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1360.063128] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1360.063293] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1360.063466] env[62066]: DEBUG nova.virt.hardware [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1360.064243] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2559dd3d-c200-49b1-9f1e-1fd4d66c8006 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.069244] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1360.069244] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5229a462-2dec-6077-cdde-f9bd2df4ad23" [ 1360.069244] env[62066]: _type = "Task" [ 1360.069244] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.076727] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5229a462-2dec-6077-cdde-f9bd2df4ad23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.579023] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5229a462-2dec-6077-cdde-f9bd2df4ad23, 'name': SearchDatastore_Task, 'duration_secs': 0.006943} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.584389] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1360.584659] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b720921-eef6-4a70-9647-0b924e92d047 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.601533] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1360.601533] env[62066]: value = "task-1156925" [ 1360.601533] env[62066]: _type = "Task" [ 1360.601533] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.608929] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156925, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.111562] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156925, 'name': ReconfigVM_Task, 'duration_secs': 0.189026} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.111946] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1361.112689] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc7e03a-d80d-43b5-83c5-692ea4685d8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.134075] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1361.134333] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d29abbf7-e505-4ef4-96ea-b14458224a63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.151054] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1361.151054] env[62066]: value = "task-1156926" [ 1361.151054] env[62066]: _type = "Task" [ 1361.151054] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.158386] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156926, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.660552] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156926, 'name': ReconfigVM_Task, 'duration_secs': 0.255402} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.660796] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Reconfigured VM instance instance-00000075 to attach disk [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16/15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1361.661640] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fa70b5-fbfc-43ae-9502-328d4d28883e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.679007] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cabdd9-c921-44fc-8003-1e059ca8213b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.695597] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020a6be4-6844-417b-852b-62e68b7b67c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.713309] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a6d34e-586d-43f5-9a62-cdef4eac5715 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.719231] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1361.719451] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecae09fa-b578-49c5-a56c-6926971b33e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.724919] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1361.724919] env[62066]: value = "task-1156927" [ 1361.724919] env[62066]: _type = "Task" [ 1361.724919] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.731696] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.235471] env[62066]: DEBUG oslo_vmware.api [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156927, 'name': PowerOnVM_Task, 'duration_secs': 0.358884} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.235862] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1363.246579] env[62066]: INFO nova.compute.manager [None req-cbf28093-4232-487e-9c91-e20d2d0f3e77 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance to original state: 'active' [ 1364.295824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.295824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.296275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1364.296275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1364.296275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1364.298424] env[62066]: INFO nova.compute.manager [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Terminating instance [ 1364.300249] env[62066]: DEBUG nova.compute.manager [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1364.300444] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1364.301301] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca8ad7d-b86e-42ef-b777-32f2e3464eb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.309222] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1364.309452] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72e5d84f-cef1-4c09-a2f6-393fa4141761 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.315483] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1364.315483] env[62066]: value = "task-1156928" [ 1364.315483] env[62066]: _type = "Task" [ 1364.315483] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.322750] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.826032] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156928, 'name': PowerOffVM_Task, 'duration_secs': 0.172734} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.826391] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1364.826572] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1364.826819] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cc8dc0c-cac1-4530-b31c-5474d606204e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.883868] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1364.884119] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1364.884294] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleting the datastore file [datastore2] 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1364.884558] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-806b0503-d94f-4e75-ae4b-bd5385159bb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.890608] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1364.890608] env[62066]: value = "task-1156930" [ 1364.890608] env[62066]: _type = "Task" [ 1364.890608] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.897876] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156930, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.400866] env[62066]: DEBUG oslo_vmware.api [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156930, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175112} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.401283] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1365.401327] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1365.401485] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1365.401694] env[62066]: INFO nova.compute.manager [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1365.401939] env[62066]: DEBUG oslo.service.loopingcall [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1365.402147] env[62066]: DEBUG nova.compute.manager [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1365.402242] env[62066]: DEBUG nova.network.neutron [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1365.839427] env[62066]: DEBUG nova.compute.manager [req-8372ad10-7a80-43c4-a4a8-e6fecc51e1df req-6baa7a16-001b-438d-8a2d-d27ac6f7b7a3 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Received event network-vif-deleted-fbffa55a-c090-4dc3-9f90-0059f6143ff7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1365.839722] env[62066]: INFO nova.compute.manager [req-8372ad10-7a80-43c4-a4a8-e6fecc51e1df req-6baa7a16-001b-438d-8a2d-d27ac6f7b7a3 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Neutron deleted interface fbffa55a-c090-4dc3-9f90-0059f6143ff7; detaching it from the instance and deleting it from the info cache [ 1365.839825] env[62066]: DEBUG nova.network.neutron [req-8372ad10-7a80-43c4-a4a8-e6fecc51e1df req-6baa7a16-001b-438d-8a2d-d27ac6f7b7a3 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1366.318737] env[62066]: DEBUG nova.network.neutron [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1366.343755] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3ff22fd-8c7a-4b0a-8218-e68878a1941b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.354463] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81de53b-96f3-420d-a1f7-a912efc9cc41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.376594] env[62066]: DEBUG nova.compute.manager [req-8372ad10-7a80-43c4-a4a8-e6fecc51e1df req-6baa7a16-001b-438d-8a2d-d27ac6f7b7a3 service nova] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Detach interface failed, port_id=fbffa55a-c090-4dc3-9f90-0059f6143ff7, reason: Instance 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1366.821398] env[62066]: INFO nova.compute.manager [-] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Took 1.42 seconds to deallocate network for instance. [ 1367.328763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1367.329161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1367.329421] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1367.351090] env[62066]: INFO nova.scheduler.client.report [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted allocations for instance 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16 [ 1367.859208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-de08fcba-b55c-4a8f-a4f1-31dafbb4feda tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.563s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1369.114997] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "9dc4a395-69a6-467d-8743-2974cd6656ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1369.115356] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1369.617885] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1369.698885] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1370.138269] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.138579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1370.140090] env[62066]: INFO nova.compute.claims [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1370.201923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.173066] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c81651d-6072-48d9-956e-a08a86017be9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.180836] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae38af5-f567-4990-9c89-a74a5c0c13cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.210331] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b546f202-d702-4b14-b506-1d9c1e485373 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.217095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172b836a-e797-42ef-8484-86fd09bb3f6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.229666] env[62066]: DEBUG nova.compute.provider_tree [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1371.732491] env[62066]: DEBUG nova.scheduler.client.report [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1372.238355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.238888] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1372.241613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.040s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.241832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.241995] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1372.243066] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f126abf-dede-43d3-8990-cb0e68b6c8a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.251523] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16924ce7-6541-45b9-8d21-7de9a97261be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.266546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f43bd9-7f3a-4a34-8f1d-caac58d2a658 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.273037] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a804cd7-0734-4633-8404-c8bbff052c2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.302640] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181470MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1372.302804] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.302990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.745474] env[62066]: DEBUG nova.compute.utils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1372.746892] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1372.747081] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1372.799757] env[62066]: DEBUG nova.policy [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd75a2f304461487e87c6d710eb7f830b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '30c8aab967844a249c72080364b29278', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1373.076148] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Successfully created port: c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1373.250941] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1373.322083] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 9dc4a395-69a6-467d-8743-2974cd6656ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1373.322294] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1373.322441] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1373.346377] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5b09ec-8516-476a-8f1d-be012c88a3f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.354452] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79332a5-f3e0-4664-b0cb-edc95a669be1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.386392] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a57393e-299a-42c0-8096-1f8165bc6f41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.393313] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5054719b-24e7-4b46-88e8-e07b6b5e4486 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.406099] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.909008] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.260191] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1374.285020] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-07T00:50:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-07T00:50:36Z,direct_url=,disk_format='vmdk',id=fc5145ed-66bc-4490-b8ac-7ca0de814dc0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a95f0d02689045adbd4d942d7a467dd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-07T00:50:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1374.285292] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1374.285458] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1374.285645] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1374.285799] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1374.285952] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1374.286179] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1374.286344] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1374.286515] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1374.286682] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1374.286856] env[62066]: DEBUG nova.virt.hardware [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1374.287771] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7592b69-fb53-407d-b5ea-d7e2df879033 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.296219] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615ae16a-9e0d-4ac3-8885-f97d614a2f55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.414431] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1374.414653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.112s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.414880] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.415030] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1374.429862] env[62066]: DEBUG nova.compute.manager [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Received event network-vif-plugged-c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1374.430097] env[62066]: DEBUG oslo_concurrency.lockutils [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] Acquiring lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1374.430312] env[62066]: DEBUG oslo_concurrency.lockutils [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1374.430496] env[62066]: DEBUG oslo_concurrency.lockutils [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.430703] env[62066]: DEBUG nova.compute.manager [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] No waiting events found dispatching network-vif-plugged-c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1374.430826] env[62066]: WARNING nova.compute.manager [req-2208ad4c-587d-4428-8c31-8156f8d3cb49 req-48430a7a-0120-4573-984d-49d61d31fa12 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Received unexpected event network-vif-plugged-c0ed85b5-6490-446f-ad04-2725a3b3ad2d for instance with vm_state building and task_state spawning. [ 1374.514741] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Successfully updated port: c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1374.921268] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] There are 15 instances to clean {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1374.921497] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 15cf1d6d-cbcb-4ce4-8b32-2ff470fd3a16] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1375.016906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1375.017087] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1375.017239] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1375.425089] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 063c6f80-d621-4d64-912d-732b480843d2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1375.548285] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1375.670870] env[62066]: DEBUG nova.network.neutron [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1375.928193] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 3eda50c7-0d3d-47d0-b450-05fb5c059b60] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.173894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1376.174282] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Instance network_info: |[{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1376.174723] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:9a:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0ed85b5-6490-446f-ad04-2725a3b3ad2d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1376.182381] env[62066]: DEBUG oslo.service.loopingcall [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1376.182570] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1376.182828] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48b8ee0f-8a6f-4ebb-88c6-3c67bfd3016b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.203023] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1376.203023] env[62066]: value = "task-1156931" [ 1376.203023] env[62066]: _type = "Task" [ 1376.203023] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.210632] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156931, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.431585] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 209fc51b-a036-4deb-9242-258f9c3e955c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.457135] env[62066]: DEBUG nova.compute.manager [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Received event network-changed-c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1376.457341] env[62066]: DEBUG nova.compute.manager [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Refreshing instance network info cache due to event network-changed-c0ed85b5-6490-446f-ad04-2725a3b3ad2d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1376.457584] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1376.457804] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1376.457993] env[62066]: DEBUG nova.network.neutron [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Refreshing network info cache for port c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1376.713312] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1156931, 'name': CreateVM_Task, 'duration_secs': 0.283358} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1376.713480] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1376.720149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1376.720376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1376.720699] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1376.720954] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32480689-c4c9-425c-a0fc-47729b70b41a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1376.725702] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1376.725702] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]5202d332-0a5e-4eb3-a1fa-e2377814dffd" [ 1376.725702] env[62066]: _type = "Task" [ 1376.725702] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1376.732989] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5202d332-0a5e-4eb3-a1fa-e2377814dffd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1376.934893] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: d139a247-e6bc-439e-b220-57cc5f135f04] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.136913] env[62066]: DEBUG nova.network.neutron [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updated VIF entry in instance network info cache for port c0ed85b5-6490-446f-ad04-2725a3b3ad2d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1377.137352] env[62066]: DEBUG nova.network.neutron [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1377.236842] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]5202d332-0a5e-4eb3-a1fa-e2377814dffd, 'name': SearchDatastore_Task, 'duration_secs': 0.011187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.237217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1377.237546] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Processing image fc5145ed-66bc-4490-b8ac-7ca0de814dc0 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1377.237861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.238119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.238391] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1377.238705] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7d33911-d2a3-40e1-baba-abc18e260348 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.246680] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1377.246923] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1377.247665] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-825c5dfe-0d06-4372-9c9c-5f4108c399a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.253932] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1377.253932] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]529a9771-5238-516d-8f04-b564ff33e2c6" [ 1377.253932] env[62066]: _type = "Task" [ 1377.253932] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.261310] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529a9771-5238-516d-8f04-b564ff33e2c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.437690] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 01135457-b646-4c30-a730-1a18468ce58a] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.639931] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae662a12-76c9-4285-91f6-dc2fcc4025fb req-0f8d9581-6fe6-4971-8ae0-58a1ae2994e5 service nova] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1377.764259] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]529a9771-5238-516d-8f04-b564ff33e2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.008312} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1377.765011] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b990aff-76f2-4329-bfbe-bf04f6c15acc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.769762] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1377.769762] env[62066]: value = "session[52597e4d-4ae7-e745-a150-d050a84ce806]52fa4553-4484-7a93-ff7f-61735f93ab25" [ 1377.769762] env[62066]: _type = "Task" [ 1377.769762] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.777430] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fa4553-4484-7a93-ff7f-61735f93ab25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.941456] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 4253e71f-4ca1-46ae-8c70-381dd9bc9383] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.281340] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': session[52597e4d-4ae7-e745-a150-d050a84ce806]52fa4553-4484-7a93-ff7f-61735f93ab25, 'name': SearchDatastore_Task, 'duration_secs': 0.009545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.281613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1378.281895] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 9dc4a395-69a6-467d-8743-2974cd6656ed/9dc4a395-69a6-467d-8743-2974cd6656ed.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1378.282165] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a17f530-b3de-4a7b-8283-88296b530356 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.288876] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1378.288876] env[62066]: value = "task-1156932" [ 1378.288876] env[62066]: _type = "Task" [ 1378.288876] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.295928] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156932, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.445815] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f17d08b0-5ac0-442e-adba-fe7c91107206] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.799182] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156932, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448859} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.799465] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fc5145ed-66bc-4490-b8ac-7ca0de814dc0/fc5145ed-66bc-4490-b8ac-7ca0de814dc0.vmdk to [datastore2] 9dc4a395-69a6-467d-8743-2974cd6656ed/9dc4a395-69a6-467d-8743-2974cd6656ed.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1378.799686] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1378.799942] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7860638a-f322-4e42-880a-5168d37d8d63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.806288] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1378.806288] env[62066]: value = "task-1156933" [ 1378.806288] env[62066]: _type = "Task" [ 1378.806288] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.813669] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156933, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.949493] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 3852b155-6bdb-466d-af8f-0e3e665fe5c8] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.316748] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156933, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249954} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.317050] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1379.317803] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9255760-d557-4c39-bef9-0f9831548a97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.338949] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 9dc4a395-69a6-467d-8743-2974cd6656ed/9dc4a395-69a6-467d-8743-2974cd6656ed.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1379.339237] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e5a05c1-64c4-4c21-a800-709e9fbbb0d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.361270] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1379.361270] env[62066]: value = "task-1156934" [ 1379.361270] env[62066]: _type = "Task" [ 1379.361270] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.369921] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.452637] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 968f6419-835e-49fe-b3a1-15cada4d22c6] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.871477] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156934, 'name': ReconfigVM_Task, 'duration_secs': 0.290785} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.871929] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 9dc4a395-69a6-467d-8743-2974cd6656ed/9dc4a395-69a6-467d-8743-2974cd6656ed.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1379.872581] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45efcd48-efed-4aa4-a80f-7b57d6b443f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.878792] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1379.878792] env[62066]: value = "task-1156935" [ 1379.878792] env[62066]: _type = "Task" [ 1379.878792] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.887039] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156935, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.955877] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: f23700a0-c060-4f7e-8a7e-7d8acaeb7762] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.389219] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156935, 'name': Rename_Task, 'duration_secs': 0.132498} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.389495] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1380.389736] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2cf4d412-697c-4420-8333-843bb0c18470 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.395246] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1380.395246] env[62066]: value = "task-1156936" [ 1380.395246] env[62066]: _type = "Task" [ 1380.395246] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.402286] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.458916] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: a451b1d1-73d3-41ad-b165-23983e48ace5] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.905246] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156936, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.961828] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 9c8bf6da-a21e-4566-9c15-4457b9a3de02] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.405833] env[62066]: DEBUG oslo_vmware.api [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156936, 'name': PowerOnVM_Task, 'duration_secs': 1.007689} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.406178] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1381.406363] env[62066]: INFO nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1381.406550] env[62066]: DEBUG nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1381.407339] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da1d81f-cbbb-45f0-b895-36b8401abc05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.464611] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 6e76c799-9341-4315-8fa9-3c9acc1f8869] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.924425] env[62066]: INFO nova.compute.manager [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Took 11.80 seconds to build instance. [ 1381.967591] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 5db50c22-048b-4cce-962a-3df1262f6e4f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1382.326146] env[62066]: DEBUG nova.compute.manager [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Received event network-changed-c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1382.326363] env[62066]: DEBUG nova.compute.manager [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Refreshing instance network info cache due to event network-changed-c0ed85b5-6490-446f-ad04-2725a3b3ad2d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1382.326582] env[62066]: DEBUG oslo_concurrency.lockutils [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1382.326761] env[62066]: DEBUG oslo_concurrency.lockutils [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1382.326898] env[62066]: DEBUG nova.network.neutron [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Refreshing network info cache for port c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1382.426278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3aaec7c3-47ba-4661-b306-fe97eb4d0637 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.311s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.470953] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.471328] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Cleaning up deleted instances with incomplete migration {{(pid=62066) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1382.973845] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.038283] env[62066]: DEBUG nova.network.neutron [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updated VIF entry in instance network info cache for port c0ed85b5-6490-446f-ad04-2725a3b3ad2d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1383.038659] env[62066]: DEBUG nova.network.neutron [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1383.541067] env[62066]: DEBUG oslo_concurrency.lockutils [req-57aa4344-3b3c-45ab-a372-3e6d13acf6cc req-f1fd3abc-189a-4e12-85bd-f4ac2cff747e service nova] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1389.478126] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.478569] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.478569] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1389.478652] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1390.008950] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.009125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.009318] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1390.009489] env[62066]: DEBUG nova.objects.instance [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lazy-loading 'info_cache' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.709318] env[62066]: DEBUG nova.network.neutron [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1392.212443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1392.212692] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1392.212858] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213032] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213207] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213369] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213511] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213653] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1392.213779] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1419.779230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "9dc4a395-69a6-467d-8743-2974cd6656ed" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1419.779608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1419.779647] env[62066]: DEBUG nova.compute.manager [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1419.780898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca626c5-16a3-456f-88dd-80f54fc71c61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.789351] env[62066]: DEBUG nova.compute.manager [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1419.789906] env[62066]: DEBUG nova.objects.instance [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'flavor' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1420.296475] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1420.296752] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b754b24-ec10-4920-b196-52b2c1ac348e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.304636] env[62066]: DEBUG oslo_vmware.api [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1420.304636] env[62066]: value = "task-1156937" [ 1420.304636] env[62066]: _type = "Task" [ 1420.304636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.313414] env[62066]: DEBUG oslo_vmware.api [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.814828] env[62066]: DEBUG oslo_vmware.api [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156937, 'name': PowerOffVM_Task, 'duration_secs': 0.18166} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.815194] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1420.815323] env[62066]: DEBUG nova.compute.manager [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1420.816076] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1d7f60-53c0-4100-9685-d4b83b12f028 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.327930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1790d675-08fe-459b-ac42-08ba59bf14ba tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.203224] env[62066]: DEBUG nova.objects.instance [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'flavor' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1422.708809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1422.709019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1422.709185] env[62066]: DEBUG nova.network.neutron [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1422.709368] env[62066]: DEBUG nova.objects.instance [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'info_cache' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1423.213323] env[62066]: DEBUG nova.objects.base [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Object Instance<9dc4a395-69a6-467d-8743-2974cd6656ed> lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1423.988020] env[62066]: DEBUG nova.network.neutron [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1424.491237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1424.994579] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1424.994911] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da266e1e-3e2d-4733-957f-0852d7c5feea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.002055] env[62066]: DEBUG oslo_vmware.api [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1425.002055] env[62066]: value = "task-1156938" [ 1425.002055] env[62066]: _type = "Task" [ 1425.002055] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1425.010728] env[62066]: DEBUG oslo_vmware.api [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.513267] env[62066]: DEBUG oslo_vmware.api [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156938, 'name': PowerOnVM_Task, 'duration_secs': 0.371475} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.513655] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1425.513740] env[62066]: DEBUG nova.compute.manager [None req-2a6c2be5-a270-4ddf-a830-9f78f2351463 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1425.514663] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038ec086-2445-424e-a0e4-c4e38a29c0ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.215667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d478db-3465-4d97-9f1b-84e77df2bcbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.222397] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1427.222622] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-45523362-6764-4bdc-a4f1-b672e63300c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.228785] env[62066]: DEBUG oslo_vmware.api [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1427.228785] env[62066]: value = "task-1156939" [ 1427.228785] env[62066]: _type = "Task" [ 1427.228785] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1427.236184] env[62066]: DEBUG oslo_vmware.api [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156939, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1427.739362] env[62066]: DEBUG oslo_vmware.api [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156939, 'name': SuspendVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.239998] env[62066]: DEBUG oslo_vmware.api [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156939, 'name': SuspendVM_Task, 'duration_secs': 0.539803} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.240392] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1428.240512] env[62066]: DEBUG nova.compute.manager [None req-e021838d-c802-4788-bd72-e43b67a9e6cf tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1428.241240] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7313829-e772-4775-9bca-a9ce180bea62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.551107] env[62066]: INFO nova.compute.manager [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Resuming [ 1429.551763] env[62066]: DEBUG nova.objects.instance [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'flavor' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1430.561021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1430.561411] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquired lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.561452] env[62066]: DEBUG nova.network.neutron [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1430.698780] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1431.202219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1431.202391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1431.202557] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1431.202710] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1431.205687] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a421ae-3ea2-4dd8-83ae-6e53a3da7a74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.213828] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1841037c-8dae-49af-98bb-ac2f2b216acb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.227937] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9498a47b-3177-4cbf-b537-b0653db8804c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.234203] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34330a0b-63cd-4fde-b110-4cb0a89c3e14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.264094] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181298MB free_disk=164GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1431.264232] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1431.264418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1431.268384] env[62066]: DEBUG nova.network.neutron [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [{"id": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "address": "fa:16:3e:c5:9a:fd", "network": {"id": "ab271bcc-712a-46ba-a8a6-352a52e2a074", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-298398749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "30c8aab967844a249c72080364b29278", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0ed85b5-64", "ovs_interfaceid": "c0ed85b5-6490-446f-ad04-2725a3b3ad2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1431.770223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Releasing lock "refresh_cache-9dc4a395-69a6-467d-8743-2974cd6656ed" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1431.771201] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b829914c-0825-411e-a461-16dc4f03e9f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.777776] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Resuming the VM {{(pid=62066) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1431.778538] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d06d1db8-13b3-4a7c-a1ac-8bafd79667cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.785431] env[62066]: DEBUG oslo_vmware.api [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1431.785431] env[62066]: value = "task-1156940" [ 1431.785431] env[62066]: _type = "Task" [ 1431.785431] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.792929] env[62066]: DEBUG oslo_vmware.api [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.296198] env[62066]: DEBUG oslo_vmware.api [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156940, 'name': PowerOnVM_Task, 'duration_secs': 0.498793} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.296504] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Resumed the VM {{(pid=62066) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1432.296691] env[62066]: DEBUG nova.compute.manager [None req-3addc874-9c57-4d98-863e-9073377db91c tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1432.297516] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ba88c1-6f34-4ed6-bac9-4e0affd67c8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.361424] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Instance 9dc4a395-69a6-467d-8743-2974cd6656ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1432.361636] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1432.361776] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1432.379016] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing inventories for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1432.393264] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating ProviderTree inventory for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1432.393464] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Updating inventory in ProviderTree for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1432.406565] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing aggregate associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1432.426174] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Refreshing trait associations for resource provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1432.454385] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2f8617-09a0-4cf3-8a9e-eaff60f7a190 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.462656] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840954cb-1793-4eeb-a54e-6458a72424ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.494187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acc259f-952c-4339-9b64-dcbe34fe08fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.502072] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20a7062-5e20-4217-99b3-7f91219b5d44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.515887] env[62066]: DEBUG nova.compute.provider_tree [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1433.019208] env[62066]: DEBUG nova.scheduler.client.report [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1433.175705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "9dc4a395-69a6-467d-8743-2974cd6656ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1433.175971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1433.176207] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1433.176397] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1433.176567] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1433.178716] env[62066]: INFO nova.compute.manager [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Terminating instance [ 1433.180468] env[62066]: DEBUG nova.compute.manager [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1433.180672] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1433.181555] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be671321-e4e4-4679-86f0-5bde201d19b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.188914] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1433.189153] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e86f5d0-0b74-4267-96d7-a2f395bee15d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.195959] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1433.195959] env[62066]: value = "task-1156941" [ 1433.195959] env[62066]: _type = "Task" [ 1433.195959] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.203483] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.524726] env[62066]: DEBUG nova.compute.resource_tracker [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1433.524921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.260s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1433.706238] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156941, 'name': PowerOffVM_Task, 'duration_secs': 0.191916} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.706504] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1433.706674] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1433.706918] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb93c182-fcb2-495b-9e86-782b6faaef82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.771344] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1433.771559] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1433.771753] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleting the datastore file [datastore2] 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1433.772036] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4adcbe2c-ae15-4890-8f22-2344a0ecd83f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.778847] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for the task: (returnval){ [ 1433.778847] env[62066]: value = "task-1156943" [ 1433.778847] env[62066]: _type = "Task" [ 1433.778847] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.786909] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.288931] env[62066]: DEBUG oslo_vmware.api [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Task: {'id': task-1156943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1414} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.289366] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1434.289451] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1434.289562] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1434.289743] env[62066]: INFO nova.compute.manager [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1434.289980] env[62066]: DEBUG oslo.service.loopingcall [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1434.290184] env[62066]: DEBUG nova.compute.manager [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1434.290276] env[62066]: DEBUG nova.network.neutron [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1434.702107] env[62066]: DEBUG nova.compute.manager [req-6b7fd636-fff0-42a6-a9fd-a4372c721536 req-67e99846-80a0-44fe-811b-875d37acc1d6 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Received event network-vif-deleted-c0ed85b5-6490-446f-ad04-2725a3b3ad2d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1434.702229] env[62066]: INFO nova.compute.manager [req-6b7fd636-fff0-42a6-a9fd-a4372c721536 req-67e99846-80a0-44fe-811b-875d37acc1d6 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Neutron deleted interface c0ed85b5-6490-446f-ad04-2725a3b3ad2d; detaching it from the instance and deleting it from the info cache [ 1434.702384] env[62066]: DEBUG nova.network.neutron [req-6b7fd636-fff0-42a6-a9fd-a4372c721536 req-67e99846-80a0-44fe-811b-875d37acc1d6 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1435.181966] env[62066]: DEBUG nova.network.neutron [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1435.204789] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-633485e0-9c2c-4a8c-8a33-52146b7966cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.214352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d711f4fa-4aee-44fc-958d-1bbf3eb6a455 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.238224] env[62066]: DEBUG nova.compute.manager [req-6b7fd636-fff0-42a6-a9fd-a4372c721536 req-67e99846-80a0-44fe-811b-875d37acc1d6 service nova] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Detach interface failed, port_id=c0ed85b5-6490-446f-ad04-2725a3b3ad2d, reason: Instance 9dc4a395-69a6-467d-8743-2974cd6656ed could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1435.685144] env[62066]: INFO nova.compute.manager [-] [instance: 9dc4a395-69a6-467d-8743-2974cd6656ed] Took 1.39 seconds to deallocate network for instance. [ 1436.191282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1436.191559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1436.191791] env[62066]: DEBUG nova.objects.instance [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lazy-loading 'resources' on Instance uuid 9dc4a395-69a6-467d-8743-2974cd6656ed {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1436.726759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864d1f78-97e8-4c00-8cb2-1784098095b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.734430] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b2111c-0aba-4f16-96ee-bec3e6e5cc47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.763618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faee6ac2-626d-4c29-a167-9db852608569 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.770959] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca38b2f2-bbc7-483a-98b1-efe1321590d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.783914] env[62066]: DEBUG nova.compute.provider_tree [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed in ProviderTree for provider: cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1437.287331] env[62066]: DEBUG nova.scheduler.client.report [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Inventory has not changed for provider cd4c0e36-9c88-4f73-a93c-1ff383ed97c4 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 164, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1437.792738] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1437.813458] env[62066]: INFO nova.scheduler.client.report [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Deleted allocations for instance 9dc4a395-69a6-467d-8743-2974cd6656ed [ 1438.322962] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e002b77a-6be8-4eee-9dc9-d6768a724377 tempest-ServerActionsTestJSON-1608519773 tempest-ServerActionsTestJSON-1608519773-project-member] Lock "9dc4a395-69a6-467d-8743-2974cd6656ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.147s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.520315] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.520575] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.025631] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.025823] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1440.025947] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1440.530090] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1440.530090] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.530090] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.530090] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.530090] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1440.530090] env[62066]: DEBUG nova.compute.manager [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1441.699056] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1441.699056] env[62066]: DEBUG oslo_service.periodic_task [None req-5c425cd5-0f44-4817-a57e-2e11530f73ce None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}